var/home/core/zuul-output/0000755000175000017500000000000015070143616014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070150124015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003624066315070150115017702 0ustar rootrootOct 04 07:09:41 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 07:09:41 crc restorecon[4672]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:41 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 07:09:42 crc restorecon[4672]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 07:09:43 crc kubenswrapper[4987]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 07:09:43 crc kubenswrapper[4987]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 07:09:43 crc kubenswrapper[4987]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 07:09:43 crc kubenswrapper[4987]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 07:09:43 crc kubenswrapper[4987]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 07:09:43 crc kubenswrapper[4987]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.824865 4987 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834067 4987 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834094 4987 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834100 4987 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834104 4987 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834108 4987 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834114 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834118 4987 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834122 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834126 4987 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834130 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834134 4987 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834138 4987 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834142 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834145 4987 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834149 4987 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834153 4987 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834156 4987 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834163 4987 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834169 4987 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834174 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834179 4987 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834184 4987 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834189 4987 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834195 4987 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834201 4987 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834205 4987 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834208 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834212 4987 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834215 4987 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834219 4987 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834224 4987 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834229 4987 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834232 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834236 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834240 4987 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834244 4987 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834248 4987 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834252 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834255 4987 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834259 4987 feature_gate.go:330] unrecognized feature gate: Example Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834263 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834267 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834271 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834275 4987 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834278 4987 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834282 4987 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834287 4987 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834291 4987 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834295 4987 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834299 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834302 4987 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834306 4987 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834310 4987 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834314 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834317 4987 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834323 4987 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834327 4987 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834330 4987 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834334 4987 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834337 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834341 4987 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834346 4987 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834350 4987 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834355 4987 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834358 4987 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834362 4987 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834366 4987 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834370 4987 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834373 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834376 4987 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.834380 4987 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834503 4987 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834518 4987 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834529 4987 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834535 4987 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834543 4987 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834548 4987 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834554 4987 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834561 4987 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834565 4987 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834570 4987 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834575 4987 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834580 4987 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834584 4987 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834589 4987 flags.go:64] FLAG: --cgroup-root="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834594 4987 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834598 4987 flags.go:64] FLAG: --client-ca-file="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834602 4987 flags.go:64] FLAG: --cloud-config="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834606 4987 flags.go:64] FLAG: --cloud-provider="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834610 4987 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834617 4987 flags.go:64] FLAG: --cluster-domain="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834653 4987 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834659 4987 flags.go:64] FLAG: --config-dir="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834664 4987 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834669 4987 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834681 4987 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834686 4987 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834691 4987 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834696 4987 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834700 4987 flags.go:64] FLAG: --contention-profiling="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834704 4987 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834709 4987 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834714 4987 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834718 4987 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834725 4987 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834729 4987 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834733 4987 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834738 4987 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834742 4987 flags.go:64] FLAG: --enable-server="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834747 4987 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834754 4987 flags.go:64] FLAG: --event-burst="100" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834758 4987 flags.go:64] FLAG: --event-qps="50" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834762 4987 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834767 4987 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834772 4987 flags.go:64] FLAG: --eviction-hard="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834777 4987 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834781 4987 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834786 4987 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834791 4987 flags.go:64] FLAG: --eviction-soft="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834796 4987 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834800 4987 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834805 4987 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834809 4987 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834814 4987 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834818 4987 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834823 4987 flags.go:64] FLAG: --feature-gates="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834828 4987 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834833 4987 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834838 4987 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834843 4987 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834848 4987 flags.go:64] FLAG: --healthz-port="10248" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834853 4987 flags.go:64] FLAG: --help="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834857 4987 flags.go:64] FLAG: --hostname-override="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834861 4987 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834866 4987 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834871 4987 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834875 4987 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834879 4987 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834883 4987 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834888 4987 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834892 4987 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834896 4987 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834902 4987 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834906 4987 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834911 4987 flags.go:64] FLAG: --kube-reserved="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834915 4987 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834920 4987 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834924 4987 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834928 4987 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834933 4987 flags.go:64] FLAG: --lock-file="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834937 4987 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834942 4987 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834947 4987 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834955 4987 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834960 4987 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834964 4987 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834969 4987 flags.go:64] FLAG: --logging-format="text" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834974 4987 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834979 4987 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834984 4987 flags.go:64] FLAG: --manifest-url="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834988 4987 flags.go:64] FLAG: --manifest-url-header="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.834998 4987 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835003 4987 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835009 4987 flags.go:64] FLAG: --max-pods="110" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835014 4987 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835018 4987 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835023 4987 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835027 4987 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835032 4987 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835036 4987 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835041 4987 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835057 4987 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835062 4987 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835066 4987 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835071 4987 flags.go:64] FLAG: --pod-cidr="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835075 4987 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835082 4987 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835086 4987 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835091 4987 flags.go:64] FLAG: --pods-per-core="0" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835095 4987 flags.go:64] FLAG: --port="10250" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835100 4987 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835105 4987 flags.go:64] FLAG: --provider-id="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835109 4987 flags.go:64] FLAG: --qos-reserved="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835114 4987 flags.go:64] FLAG: --read-only-port="10255" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835118 4987 flags.go:64] FLAG: --register-node="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835122 4987 flags.go:64] FLAG: --register-schedulable="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835127 4987 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835136 4987 flags.go:64] FLAG: --registry-burst="10" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835141 4987 flags.go:64] FLAG: --registry-qps="5" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835145 4987 flags.go:64] FLAG: --reserved-cpus="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835150 4987 flags.go:64] FLAG: --reserved-memory="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835156 4987 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835161 4987 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835165 4987 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835170 4987 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835174 4987 flags.go:64] FLAG: --runonce="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835179 4987 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835183 4987 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835187 4987 flags.go:64] FLAG: --seccomp-default="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835192 4987 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835196 4987 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835201 4987 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835206 4987 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835210 4987 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835214 4987 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835219 4987 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835231 4987 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835235 4987 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835240 4987 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835244 4987 flags.go:64] FLAG: --system-cgroups="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835248 4987 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835256 4987 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835260 4987 flags.go:64] FLAG: --tls-cert-file="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835265 4987 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835271 4987 flags.go:64] FLAG: --tls-min-version="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835276 4987 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835280 4987 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835284 4987 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835289 4987 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835294 4987 flags.go:64] FLAG: --v="2" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835300 4987 flags.go:64] FLAG: --version="false" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835307 4987 flags.go:64] FLAG: --vmodule="" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835312 4987 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835317 4987 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835435 4987 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835440 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835444 4987 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835449 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835453 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835457 4987 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835460 4987 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835464 4987 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835467 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835470 4987 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835474 4987 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835477 4987 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835481 4987 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835484 4987 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835490 4987 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835495 4987 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835499 4987 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835503 4987 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835507 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835511 4987 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835515 4987 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835518 4987 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835522 4987 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835525 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835530 4987 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835534 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835539 4987 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835542 4987 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835546 4987 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835549 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835553 4987 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835556 4987 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835560 4987 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835563 4987 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835570 4987 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835573 4987 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835577 4987 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835580 4987 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835585 4987 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835588 4987 feature_gate.go:330] unrecognized feature gate: Example Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835592 4987 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835596 4987 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835600 4987 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835603 4987 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835607 4987 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835610 4987 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835615 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835631 4987 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835635 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835639 4987 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835642 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835646 4987 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835649 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835653 4987 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835656 4987 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835659 4987 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835663 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835666 4987 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835670 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835674 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835677 4987 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835681 4987 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835684 4987 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835688 4987 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835691 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835695 4987 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835700 4987 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835703 4987 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835707 4987 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835710 4987 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.835713 4987 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.835726 4987 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.847440 4987 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.847488 4987 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847616 4987 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847658 4987 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847664 4987 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847671 4987 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847678 4987 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847685 4987 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847690 4987 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847695 4987 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847701 4987 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847707 4987 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847712 4987 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847717 4987 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847723 4987 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847728 4987 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847734 4987 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847739 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847746 4987 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847752 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847757 4987 feature_gate.go:330] unrecognized feature gate: Example Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847763 4987 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847769 4987 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847774 4987 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847779 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847785 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847790 4987 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847797 4987 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847802 4987 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847808 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847814 4987 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847820 4987 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847827 4987 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847833 4987 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847839 4987 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847846 4987 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847855 4987 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847861 4987 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847866 4987 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847872 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847877 4987 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847882 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847888 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847893 4987 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847898 4987 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847903 4987 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847908 4987 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847914 4987 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847919 4987 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847924 4987 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847929 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847934 4987 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847941 4987 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847948 4987 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847954 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847960 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847965 4987 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847972 4987 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847978 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847983 4987 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847990 4987 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.847996 4987 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848002 4987 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848007 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848013 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848018 4987 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848023 4987 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848029 4987 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848034 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848040 4987 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848046 4987 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848053 4987 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848060 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.848070 4987 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848296 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848305 4987 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848311 4987 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848318 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848324 4987 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848329 4987 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848335 4987 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848341 4987 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848347 4987 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848353 4987 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848360 4987 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848365 4987 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848372 4987 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848377 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848382 4987 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848388 4987 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848393 4987 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848398 4987 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848404 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848409 4987 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848414 4987 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848420 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848425 4987 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848472 4987 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848479 4987 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848484 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848490 4987 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848496 4987 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848502 4987 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848509 4987 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848514 4987 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848522 4987 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848530 4987 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848535 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848543 4987 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848549 4987 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848555 4987 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848562 4987 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848568 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848573 4987 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848578 4987 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848584 4987 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848590 4987 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848595 4987 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848600 4987 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848606 4987 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848611 4987 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848616 4987 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848639 4987 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848645 4987 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848650 4987 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848655 4987 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848661 4987 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848666 4987 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848671 4987 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848676 4987 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848681 4987 feature_gate.go:330] unrecognized feature gate: Example Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848687 4987 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848692 4987 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848699 4987 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848706 4987 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848743 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848750 4987 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848755 4987 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848761 4987 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848767 4987 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848773 4987 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848779 4987 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848786 4987 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848792 4987 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.848799 4987 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.848808 4987 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.849024 4987 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.853618 4987 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.853743 4987 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.857125 4987 server.go:997] "Starting client certificate rotation" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.857170 4987 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.857406 4987 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-06 00:37:46.742778357 +0000 UTC Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.857659 4987 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2249h28m2.885123085s for next certificate rotation Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.882026 4987 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.883688 4987 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.902572 4987 log.go:25] "Validated CRI v1 runtime API" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.935579 4987 log.go:25] "Validated CRI v1 image API" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.937679 4987 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.946471 4987 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-07-06-00-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.946514 4987 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.968106 4987 manager.go:217] Machine: {Timestamp:2025-10-04 07:09:43.964355209 +0000 UTC m=+1.019254071 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799886 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7dabc7f6-4527-406f-a679-876deca45080 BootID:4ae34cc5-3302-4911-8938-ea69ea878ab9 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c0:24:02 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c0:24:02 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9d:d1:bf Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:43:23:41 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e9:ab:05 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:1c:dc:1c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:62:ed:31:c7:e3:8d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:74:8f:ff:38:e7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.968365 4987 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.968562 4987 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.968987 4987 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.969169 4987 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.969206 4987 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.969432 4987 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.969440 4987 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.970012 4987 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.970779 4987 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.971918 4987 state_mem.go:36] "Initialized new in-memory state store" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.972023 4987 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.978458 4987 kubelet.go:418] "Attempting to sync node with API server" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.978494 4987 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.978560 4987 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.978579 4987 kubelet.go:324] "Adding apiserver pod source" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.978600 4987 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.984564 4987 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.986016 4987 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.986184 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:43 crc kubenswrapper[4987]: W1004 07:09:43.986267 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:43 crc kubenswrapper[4987]: E1004 07:09:43.986380 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:43 crc kubenswrapper[4987]: E1004 07:09:43.986459 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.988663 4987 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990288 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990321 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990330 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990338 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990384 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990396 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990406 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990420 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990432 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990441 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990470 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.990480 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.991263 4987 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.992184 4987 server.go:1280] "Started kubelet" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.992409 4987 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.992808 4987 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.992936 4987 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.993365 4987 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.994980 4987 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.995037 4987 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.995168 4987 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 07:48:25.845077014 +0000 UTC Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.995251 4987 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1584h38m41.849829977s for next certificate rotation Oct 04 07:09:43 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.996519 4987 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.996538 4987 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 07:09:43 crc kubenswrapper[4987]: E1004 07:09:43.998375 4987 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 07:09:43 crc kubenswrapper[4987]: I1004 07:09:43.998698 4987 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:43.999081 4987 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.86:6443: connect: connection refused" interval="200ms" Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.000715 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.001048 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.002486 4987 factory.go:55] Registering systemd factory Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.002512 4987 factory.go:221] Registration of the systemd container factory successfully Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.002962 4987 factory.go:153] Registering CRI-O factory Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.003009 4987 factory.go:221] Registration of the crio container factory successfully Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.003194 4987 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.003318 4987 factory.go:103] Registering Raw factory Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.005596 4987 manager.go:1196] Started watching for new ooms in manager Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.002466 4987 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.86:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b381eb734488d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 07:09:43.992141965 +0000 UTC m=+1.047040827,LastTimestamp:2025-10-04 07:09:43.992141965 +0000 UTC m=+1.047040827,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.006642 4987 server.go:460] "Adding debug handlers to kubelet server" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.008599 4987 manager.go:319] Starting recovery of all containers Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015377 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015506 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015541 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015573 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015590 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015605 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015694 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015751 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015792 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015809 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015883 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015916 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015932 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.015960 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016016 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016047 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016109 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016141 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016158 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016198 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016213 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016243 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016271 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016299 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016313 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016327 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016363 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016378 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016392 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016406 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016421 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016453 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016485 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016524 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016540 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016569 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016600 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016702 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016735 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016748 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016760 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016774 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016787 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016799 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016818 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016833 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016904 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016918 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016950 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016966 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.016979 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017004 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017034 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017051 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017077 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017091 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017116 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017130 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017142 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017154 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017167 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017193 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017205 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017217 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017244 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017258 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017272 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017301 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017317 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017352 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017381 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017397 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017427 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017441 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017471 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017486 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.017500 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018074 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018124 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018142 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018161 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018177 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018192 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018205 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018217 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018233 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018249 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018263 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018279 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018291 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018304 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018318 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018331 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018343 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018357 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018373 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018387 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.018399 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019128 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019157 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019183 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019200 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019217 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019236 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019269 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019294 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019316 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019336 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019357 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019378 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019396 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019414 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019434 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019452 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019467 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019482 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019502 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019517 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019531 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019549 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019566 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019581 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019597 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019611 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019649 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019665 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019679 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019708 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019724 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019738 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019753 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019768 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019787 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019803 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019817 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019831 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019845 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019859 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019875 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019889 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019902 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019916 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019930 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019944 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019961 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019976 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.019989 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020003 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020019 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020032 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020047 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020061 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020074 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020091 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020106 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020119 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020132 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020148 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020161 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020201 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020221 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020234 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020249 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020263 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020278 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020293 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020308 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020322 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020336 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020351 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020364 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020379 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020394 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020407 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020421 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020442 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020458 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020477 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020493 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020508 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020523 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020539 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020553 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020568 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020583 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020599 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020658 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020674 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020692 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020712 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.020727 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023149 4987 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023449 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023466 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023480 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023495 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023509 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023523 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023538 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023553 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023565 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023580 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023594 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023608 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023653 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023671 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023716 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023730 4987 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023742 4987 reconstruct.go:97] "Volume reconstruction finished" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.023751 4987 reconciler.go:26] "Reconciler: start to sync state" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.026799 4987 manager.go:324] Recovery completed Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.040496 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.042380 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.042425 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.042436 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.047226 4987 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.047250 4987 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.047280 4987 state_mem.go:36] "Initialized new in-memory state store" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.049671 4987 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.052117 4987 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.052175 4987 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.052231 4987 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.052291 4987 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.054393 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.054462 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.060383 4987 policy_none.go:49] "None policy: Start" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.061640 4987 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.061678 4987 state_mem.go:35] "Initializing new in-memory state store" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.099037 4987 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.116233 4987 manager.go:334] "Starting Device Plugin manager" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.116350 4987 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.116373 4987 server.go:79] "Starting device plugin registration server" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.117045 4987 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.117069 4987 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.118216 4987 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.118546 4987 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.118642 4987 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.124361 4987 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.153090 4987 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.153273 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.155082 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.155182 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.155206 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.155544 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.155738 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.155777 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159380 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159412 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159423 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159418 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159530 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159799 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159947 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.159980 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.160921 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.160942 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.160951 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.161008 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.161026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.161035 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.161193 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.161511 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.161566 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.162134 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.162158 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.162169 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.162301 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.162480 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.162537 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.163186 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.163224 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.163239 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164290 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164304 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164612 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164651 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164666 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164841 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.164873 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.165426 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.165452 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.165464 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.200043 4987 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.86:6443: connect: connection refused" interval="400ms" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.217854 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.219221 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.219262 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.219274 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.219303 4987 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.219808 4987 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.86:6443: connect: connection refused" node="crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225122 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225158 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225178 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225198 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225218 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225238 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225261 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225284 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225341 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225399 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225418 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225442 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225514 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225579 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.225598 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327030 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327110 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327139 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327155 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327172 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327197 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327214 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327230 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327253 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327269 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327287 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327302 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327330 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327361 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327360 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327440 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327483 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327493 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327530 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327448 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327397 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327388 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327385 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327589 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327498 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327562 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327391 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327563 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327704 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.327550 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.420688 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.423334 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.423388 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.423400 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.423432 4987 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.424131 4987 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.86:6443: connect: connection refused" node="crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.503059 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.509332 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.526944 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.549722 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.555999 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.601747 4987 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.86:6443: connect: connection refused" interval="800ms" Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.611942 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-531ef3213cad1cb133484a70c1260dc07aa7b569aec8fd80b8baa105b0052f3a WatchSource:0}: Error finding container 531ef3213cad1cb133484a70c1260dc07aa7b569aec8fd80b8baa105b0052f3a: Status 404 returned error can't find the container with id 531ef3213cad1cb133484a70c1260dc07aa7b569aec8fd80b8baa105b0052f3a Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.613716 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f15847b5d9ff24330a3ae97cd336a59b4924f40d3e6e38e1c4fb408b792d5174 WatchSource:0}: Error finding container f15847b5d9ff24330a3ae97cd336a59b4924f40d3e6e38e1c4fb408b792d5174: Status 404 returned error can't find the container with id f15847b5d9ff24330a3ae97cd336a59b4924f40d3e6e38e1c4fb408b792d5174 Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.618374 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-892220669cc5764cab7d38215d0a3da6e64a609f2c61da135f3b8a0880b4edc0 WatchSource:0}: Error finding container 892220669cc5764cab7d38215d0a3da6e64a609f2c61da135f3b8a0880b4edc0: Status 404 returned error can't find the container with id 892220669cc5764cab7d38215d0a3da6e64a609f2c61da135f3b8a0880b4edc0 Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.619026 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-685d328d6cd3b4d17d0e8fc929fa829b4e33b6854aa1b824d8f4c582fdebbd93 WatchSource:0}: Error finding container 685d328d6cd3b4d17d0e8fc929fa829b4e33b6854aa1b824d8f4c582fdebbd93: Status 404 returned error can't find the container with id 685d328d6cd3b4d17d0e8fc929fa829b4e33b6854aa1b824d8f4c582fdebbd93 Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.619953 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-cb0edb7f4b4e1fd40de060ab201d4792ac88dbf27e0e577c02c9b8c0e5213f26 WatchSource:0}: Error finding container cb0edb7f4b4e1fd40de060ab201d4792ac88dbf27e0e577c02c9b8c0e5213f26: Status 404 returned error can't find the container with id cb0edb7f4b4e1fd40de060ab201d4792ac88dbf27e0e577c02c9b8c0e5213f26 Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.824364 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.826259 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.826330 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.826347 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.826388 4987 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.827112 4987 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.86:6443: connect: connection refused" node="crc" Oct 04 07:09:44 crc kubenswrapper[4987]: W1004 07:09:44.926591 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:44 crc kubenswrapper[4987]: E1004 07:09:44.926753 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:44 crc kubenswrapper[4987]: I1004 07:09:44.993992 4987 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.060385 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"685d328d6cd3b4d17d0e8fc929fa829b4e33b6854aa1b824d8f4c582fdebbd93"} Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.062150 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"892220669cc5764cab7d38215d0a3da6e64a609f2c61da135f3b8a0880b4edc0"} Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.063869 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f15847b5d9ff24330a3ae97cd336a59b4924f40d3e6e38e1c4fb408b792d5174"} Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.064993 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"531ef3213cad1cb133484a70c1260dc07aa7b569aec8fd80b8baa105b0052f3a"} Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.066154 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cb0edb7f4b4e1fd40de060ab201d4792ac88dbf27e0e577c02c9b8c0e5213f26"} Oct 04 07:09:45 crc kubenswrapper[4987]: W1004 07:09:45.145447 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:45 crc kubenswrapper[4987]: E1004 07:09:45.145536 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:45 crc kubenswrapper[4987]: W1004 07:09:45.267582 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:45 crc kubenswrapper[4987]: E1004 07:09:45.268135 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:45 crc kubenswrapper[4987]: W1004 07:09:45.271927 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:45 crc kubenswrapper[4987]: E1004 07:09:45.272059 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:45 crc kubenswrapper[4987]: E1004 07:09:45.403090 4987 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.86:6443: connect: connection refused" interval="1.6s" Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.627984 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.631020 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.631112 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.631129 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.631174 4987 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 07:09:45 crc kubenswrapper[4987]: E1004 07:09:45.631932 4987 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.86:6443: connect: connection refused" node="crc" Oct 04 07:09:45 crc kubenswrapper[4987]: I1004 07:09:45.993451 4987 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.070709 4987 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f" exitCode=0 Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.070778 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.070890 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.072440 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.072470 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.072480 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.073386 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.073414 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.073425 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.073435 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.073539 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.074525 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.074547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.074559 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.076565 4987 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e" exitCode=0 Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.076633 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.076671 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.077593 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.077617 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.077637 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.078716 4987 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12" exitCode=0 Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.078764 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.078848 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.079479 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.079504 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.079514 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081027 4987 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d9d4f0e733e8e183d447b0dd47effd78f97e90490604b1931970e0cdff442849" exitCode=0 Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081060 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081090 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d9d4f0e733e8e183d447b0dd47effd78f97e90490604b1931970e0cdff442849"} Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081155 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081694 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081733 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081751 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081813 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081831 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.081841 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:46 crc kubenswrapper[4987]: W1004 07:09:46.975522 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:46 crc kubenswrapper[4987]: E1004 07:09:46.975643 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:46 crc kubenswrapper[4987]: I1004 07:09:46.994241 4987 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:47 crc kubenswrapper[4987]: E1004 07:09:47.004439 4987 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.86:6443: connect: connection refused" interval="3.2s" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.086357 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.086411 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.086425 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.086537 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.087546 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.087569 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.087582 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.091301 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.091329 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.091344 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.091357 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.091370 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.091412 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.092385 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.092420 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.092432 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.093125 4987 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="995ad9437c8b74a383c41fb3a406dd528ea85a3e23e0d98bfe0bdead4da27f45" exitCode=0 Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.093197 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.093198 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"995ad9437c8b74a383c41fb3a406dd528ea85a3e23e0d98bfe0bdead4da27f45"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.093832 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.093866 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.093879 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.094777 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22"} Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.094804 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.094828 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.095526 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.095557 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.095528 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.095572 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.095589 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.095604 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.232997 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.234129 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.234162 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.234175 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.234199 4987 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 07:09:47 crc kubenswrapper[4987]: E1004 07:09:47.234643 4987 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.86:6443: connect: connection refused" node="crc" Oct 04 07:09:47 crc kubenswrapper[4987]: I1004 07:09:47.319024 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:09:47 crc kubenswrapper[4987]: W1004 07:09:47.568685 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.86:6443: connect: connection refused Oct 04 07:09:47 crc kubenswrapper[4987]: E1004 07:09:47.568771 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.86:6443: connect: connection refused" logger="UnhandledError" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100020 4987 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ab22b383b069ba13f88901506de73701628021a458d46ce9d49c8678bc38a091" exitCode=0 Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100096 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ab22b383b069ba13f88901506de73701628021a458d46ce9d49c8678bc38a091"} Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100139 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100213 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100250 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100160 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.100214 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.101505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.101535 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.101545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.101743 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.101765 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.101776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.102062 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.102095 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.102104 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.102105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.102210 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:48 crc kubenswrapper[4987]: I1004 07:09:48.102231 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.106864 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.106857 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"35c8aaec5f9ef6e5f3b19b071b695d35670cc0538b6c98da2a77bb449505e278"} Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107203 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58145476f65cbef6b951bc7b69ef33426ca11a023c88f1d12b82e740a1840190"} Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107242 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f3d74e8eb5ae1ca1a680a33beed1d4a467226145e06b7acd6613fdefab6ebfc3"} Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107255 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b962efbb367f878e7fa419ab22cae2aab7e2e6ad71f09173d537aaa5b709d884"} Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107267 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"53d95462c8244cf0dba5c28cc76b342c450f994c2a04077ecab7e3f0429a3823"} Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.106958 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107907 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107941 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.107950 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.108599 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.108647 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.108662 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.916884 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.917100 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.917161 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.918594 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.918660 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:49 crc kubenswrapper[4987]: I1004 07:09:49.918674 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.110331 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.111749 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.111795 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.111806 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.435790 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.438004 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.438157 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.438185 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:50 crc kubenswrapper[4987]: I1004 07:09:50.438239 4987 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.552245 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.552486 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.554026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.554068 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.554087 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.556724 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.748878 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.749312 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.749417 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.751287 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.751321 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.751331 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:51 crc kubenswrapper[4987]: I1004 07:09:51.921164 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.115311 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.116556 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.116606 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.116620 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.294790 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.629939 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.630207 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.631775 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.631824 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:52 crc kubenswrapper[4987]: I1004 07:09:52.631835 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.118580 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.118702 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.120177 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.120256 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.120283 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.167488 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.167714 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.169911 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.170041 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:53 crc kubenswrapper[4987]: I1004 07:09:53.170057 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.120894 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.120953 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.121959 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.121999 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.122015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:54 crc kubenswrapper[4987]: E1004 07:09:54.124471 4987 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.376769 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.921601 4987 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 07:09:54 crc kubenswrapper[4987]: I1004 07:09:54.921787 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 07:09:55 crc kubenswrapper[4987]: I1004 07:09:55.123192 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:55 crc kubenswrapper[4987]: I1004 07:09:55.124306 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:55 crc kubenswrapper[4987]: I1004 07:09:55.124343 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:55 crc kubenswrapper[4987]: I1004 07:09:55.124352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:55 crc kubenswrapper[4987]: I1004 07:09:55.129189 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:09:56 crc kubenswrapper[4987]: I1004 07:09:56.126062 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:56 crc kubenswrapper[4987]: I1004 07:09:56.126962 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:56 crc kubenswrapper[4987]: I1004 07:09:56.127003 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:56 crc kubenswrapper[4987]: I1004 07:09:56.127019 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:57 crc kubenswrapper[4987]: W1004 07:09:57.594932 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 04 07:09:57 crc kubenswrapper[4987]: I1004 07:09:57.595068 4987 trace.go:236] Trace[437316030]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 07:09:47.593) (total time: 10001ms): Oct 04 07:09:57 crc kubenswrapper[4987]: Trace[437316030]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:09:57.594) Oct 04 07:09:57 crc kubenswrapper[4987]: Trace[437316030]: [10.001826966s] [10.001826966s] END Oct 04 07:09:57 crc kubenswrapper[4987]: E1004 07:09:57.595106 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 04 07:09:57 crc kubenswrapper[4987]: I1004 07:09:57.805870 4987 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35048->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 04 07:09:57 crc kubenswrapper[4987]: I1004 07:09:57.805980 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:35048->192.168.126.11:17697: read: connection reset by peer" Oct 04 07:09:57 crc kubenswrapper[4987]: I1004 07:09:57.993960 4987 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 07:09:58 crc kubenswrapper[4987]: W1004 07:09:58.068603 4987 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.068816 4987 trace.go:236] Trace[147761100]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 07:09:48.066) (total time: 10002ms): Oct 04 07:09:58 crc kubenswrapper[4987]: Trace[147761100]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:09:58.068) Oct 04 07:09:58 crc kubenswrapper[4987]: Trace[147761100]: [10.002190117s] [10.002190117s] END Oct 04 07:09:58 crc kubenswrapper[4987]: E1004 07:09:58.068853 4987 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.352168 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.352530 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.354254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.354300 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.354313 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:58 crc kubenswrapper[4987]: I1004 07:09:58.515699 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.137102 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.139206 4987 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845" exitCode=255 Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.139316 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845"} Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.139444 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.139618 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.140719 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.140767 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.140782 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.144417 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.144485 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.144505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.145546 4987 scope.go:117] "RemoveContainer" containerID="410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.154952 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.596363 4987 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.596482 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.603033 4987 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 04 07:09:59 crc kubenswrapper[4987]: I1004 07:09:59.603113 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.143931 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.145691 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea"} Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.145789 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.145870 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.146727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.146753 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.146763 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.146801 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.146826 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:00 crc kubenswrapper[4987]: I1004 07:10:00.146835 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.753972 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.754123 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.754232 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.755443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.755487 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.755503 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:01 crc kubenswrapper[4987]: I1004 07:10:01.758792 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:10:02 crc kubenswrapper[4987]: I1004 07:10:02.152697 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:10:02 crc kubenswrapper[4987]: I1004 07:10:02.154071 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:02 crc kubenswrapper[4987]: I1004 07:10:02.154098 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:02 crc kubenswrapper[4987]: I1004 07:10:02.154107 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:03 crc kubenswrapper[4987]: I1004 07:10:03.155834 4987 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 07:10:03 crc kubenswrapper[4987]: I1004 07:10:03.157080 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:03 crc kubenswrapper[4987]: I1004 07:10:03.157140 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:03 crc kubenswrapper[4987]: I1004 07:10:03.157163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:03 crc kubenswrapper[4987]: I1004 07:10:03.300978 4987 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.124565 4987 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.429354 4987 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.599973 4987 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.603581 4987 trace.go:236] Trace[510145834]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 07:09:52.202) (total time: 12400ms): Oct 04 07:10:04 crc kubenswrapper[4987]: Trace[510145834]: ---"Objects listed" error: 12400ms (07:10:04.603) Oct 04 07:10:04 crc kubenswrapper[4987]: Trace[510145834]: [12.400737451s] [12.400737451s] END Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.603609 4987 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.603830 4987 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.606021 4987 trace.go:236] Trace[2126670323]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 07:09:53.172) (total time: 11433ms): Oct 04 07:10:04 crc kubenswrapper[4987]: Trace[2126670323]: ---"Objects listed" error: 11433ms (07:10:04.605) Oct 04 07:10:04 crc kubenswrapper[4987]: Trace[2126670323]: [11.43328962s] [11.43328962s] END Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.606044 4987 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.613205 4987 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.613341 4987 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.614724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.614839 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.614920 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.614998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.615068 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.626869 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.630747 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.630782 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.630791 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.630808 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.630817 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.640443 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.644505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.644544 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.644565 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.644570 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.644580 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.644692 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.652746 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.658296 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.661406 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.661444 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.661454 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.661470 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.661482 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.672937 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.676597 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.676669 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.676685 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.676706 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.676718 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.687459 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.687573 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.689822 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.689869 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.689883 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.689905 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.689917 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.791696 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.791730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.791740 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.791754 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.791765 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.893382 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.893414 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.893422 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.893437 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.893445 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.993957 4987 apiserver.go:52] "Watching apiserver" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.995988 4987 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.995994 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996090 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996108 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996123 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996133 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:04Z","lastTransitionTime":"2025-10-04T07:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996178 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996475 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996532 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996557 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996588 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.996648 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996666 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.996687 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.996736 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:04 crc kubenswrapper[4987]: E1004 07:10:04.996656 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:04 crc kubenswrapper[4987]: I1004 07:10:04.999672 4987 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.002122 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.002234 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.002353 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.002589 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.003377 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.003457 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.003410 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.003498 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.003427 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.005896 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.005929 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.005949 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.005969 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.005989 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006011 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006027 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006042 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006060 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006078 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006094 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006114 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006161 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006181 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006204 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006222 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006239 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006260 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006275 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006292 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006295 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006319 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006326 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006308 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006448 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006477 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006498 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006521 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006522 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006535 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006557 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006604 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006640 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006658 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006679 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006703 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006725 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006746 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006769 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006790 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006794 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006790 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006812 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006852 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006879 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006903 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006929 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006951 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006976 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006999 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007025 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007049 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007145 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007175 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007202 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007225 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007246 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007268 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007291 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007316 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007338 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007360 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007384 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007409 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007434 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007457 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007482 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007509 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007532 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007557 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007580 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007601 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007643 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007737 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007761 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007784 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007810 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007837 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007863 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007884 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007909 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007931 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007996 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008026 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008051 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008074 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008099 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008122 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008144 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008166 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008188 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008214 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008241 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008267 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008296 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008320 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008348 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008371 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008394 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008421 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008445 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008469 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008495 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008518 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008543 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008566 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008586 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008683 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008714 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008738 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008760 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008782 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008970 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008989 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009025 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009042 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009058 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009078 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009102 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009123 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009140 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009156 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009171 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009187 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009203 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009219 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009235 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009252 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009271 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009289 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009305 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009323 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009340 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009356 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009372 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009393 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009417 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009437 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009457 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009474 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009496 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009521 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009545 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009568 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009590 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009616 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009661 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009687 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009714 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009740 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009769 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009795 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009820 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009838 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009881 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009907 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009937 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009963 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009981 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010006 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006801 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006856 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006930 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.006951 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007034 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007070 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007108 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007169 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007154 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007190 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007257 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014519 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007271 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007360 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007434 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007651 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007668 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007669 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007783 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007803 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007848 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.007882 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008037 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008149 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008197 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008238 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008250 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008244 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008433 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008434 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008441 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008463 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008468 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008637 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008650 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008655 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.008708 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009585 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009599 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.009905 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.010032 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:10:05.510012982 +0000 UTC m=+22.564911824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010105 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010210 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010270 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010376 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010514 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010529 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010606 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010727 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010976 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011076 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.010988 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011149 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011264 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011331 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011416 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011463 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011736 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.011738 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012022 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012016 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012064 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012110 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012243 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012315 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012346 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012425 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012502 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012874 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012899 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.012968 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013069 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013166 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013255 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013267 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013507 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013527 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013538 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013808 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013822 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.013844 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014105 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014880 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014122 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014142 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014154 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014397 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014374 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.014783 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015023 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015053 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015080 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015098 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015114 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015132 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015150 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015151 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015167 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015151 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015184 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015202 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015219 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015237 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015265 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015283 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015297 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015356 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015394 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015418 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015419 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015443 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015455 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015468 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015496 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015519 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015543 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015568 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015593 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015616 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015654 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015676 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015702 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015725 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015748 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015770 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015792 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015814 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015833 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015850 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015867 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015885 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015900 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015917 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015935 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015954 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.015971 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016014 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016035 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016053 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016075 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016092 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016113 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016129 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016147 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016163 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016185 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016202 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016219 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016238 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016259 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016314 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016325 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016335 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016344 4987 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016353 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016363 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016373 4987 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016383 4987 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016393 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016403 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016413 4987 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016422 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016432 4987 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016441 4987 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016453 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016464 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016474 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016483 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016495 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016505 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016514 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016522 4987 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016532 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016541 4987 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016550 4987 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016559 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016568 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016581 4987 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016592 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016601 4987 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016610 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016619 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016651 4987 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016662 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016671 4987 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016682 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016690 4987 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016700 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016710 4987 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016721 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016732 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016741 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016754 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016764 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016773 4987 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016785 4987 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016795 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.016805 4987 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.017206 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.017532 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.017746 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.017838 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.018250 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.018453 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.018619 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.018896 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.019105 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.019408 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.019727 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.021573 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.021948 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.022112 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.022265 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.022922 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.023104 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.024108 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.024277 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.024699 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.025148 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.025388 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.026815 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.027059 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.027604 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.027836 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.027841 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.028316 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.028386 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.028408 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029155 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029222 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029103 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029349 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029417 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029573 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.029671 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030129 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030326 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030357 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030042 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030721 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030768 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030827 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030948 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.030858 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031058 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031376 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031571 4987 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031581 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031574 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031641 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031692 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.031742 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.033480 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:05.533464369 +0000 UTC m=+22.588363211 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.031906 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.033600 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.031911 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032142 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032162 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032361 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032392 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032540 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032512 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032658 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032761 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032834 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032970 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.032985 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.033052 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.033005 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.033270 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.033359 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.033949 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.034033 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.034437 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:05.534415217 +0000 UTC m=+22.589314139 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.034693 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.034966 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.035016 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.035079 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.035304 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.035388 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.036457 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.036583 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.036617 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.037485 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.038366 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.039159 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.039266 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.039963 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.042215 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.048430 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.048461 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.048476 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.048545 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:05.5485241 +0000 UTC m=+22.603422942 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.050087 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.051446 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.052367 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.053206 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.055307 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.055437 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.061009 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.061993 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.062047 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.062318 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.062337 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.062642 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.062939 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:05.562914171 +0000 UTC m=+22.617813013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.063774 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.063832 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.068733 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.069101 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.071309 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.073786 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.081702 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.085135 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.099340 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.099386 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.099399 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.099419 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.099431 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.100954 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.114838 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.117434 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.117610 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.118990 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119109 4987 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119170 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119234 4987 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119294 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119357 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119466 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119525 4987 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119581 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119656 4987 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119720 4987 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119775 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.119834 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.120779 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.121383 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122185 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122291 4987 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122364 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122498 4987 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122583 4987 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122675 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122744 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122818 4987 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.122933 4987 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123011 4987 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123081 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123145 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123200 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123258 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123325 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123396 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123463 4987 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123521 4987 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123573 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123651 4987 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123728 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123788 4987 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123844 4987 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123896 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.123955 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124084 4987 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124146 4987 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124200 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124258 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124311 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124389 4987 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124484 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124564 4987 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124682 4987 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124757 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124890 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124952 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.124979 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125001 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125023 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125038 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125051 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125068 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125086 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125099 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125113 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125128 4987 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125147 4987 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125165 4987 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125202 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125217 4987 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125231 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125248 4987 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125260 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125274 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125287 4987 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125306 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125319 4987 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125331 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125346 4987 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125361 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125373 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125387 4987 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125406 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125420 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125432 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125445 4987 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125460 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125472 4987 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125484 4987 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125496 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125511 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125523 4987 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125534 4987 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125546 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125561 4987 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125573 4987 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125584 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125603 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125618 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125649 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125662 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125679 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125692 4987 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125711 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125726 4987 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125743 4987 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125756 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125770 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125788 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125820 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125834 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125848 4987 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125921 4987 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.125966 4987 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126683 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126713 4987 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126727 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126740 4987 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126753 4987 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126767 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126780 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126791 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126803 4987 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126816 4987 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126830 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126842 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126856 4987 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126867 4987 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126879 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126891 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126903 4987 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126914 4987 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126926 4987 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126938 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126950 4987 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126963 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126978 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.126989 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127001 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127013 4987 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127025 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127036 4987 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127048 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127061 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127074 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127087 4987 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127101 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.127115 4987 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.131921 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.137062 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.140152 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.141269 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.150368 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.174106 4987 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.174727 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.201577 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.201633 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.201643 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.201658 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.201686 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.219542 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.222403 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.228073 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.228111 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.228124 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.228137 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.303987 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.304038 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.304048 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.304067 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.304079 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.314185 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.324586 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 07:10:05 crc kubenswrapper[4987]: W1004 07:10:05.326996 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ef3d8bac8e6d80b56ae09f5267982b0a56a9923e42f5b919a62ccd241c5f3fe1 WatchSource:0}: Error finding container ef3d8bac8e6d80b56ae09f5267982b0a56a9923e42f5b919a62ccd241c5f3fe1: Status 404 returned error can't find the container with id ef3d8bac8e6d80b56ae09f5267982b0a56a9923e42f5b919a62ccd241c5f3fe1 Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.333162 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 07:10:05 crc kubenswrapper[4987]: W1004 07:10:05.334548 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-b915051f7799a8520fc50f755ae42139e26ed7cd7e74df153d3792e4f8681e6d WatchSource:0}: Error finding container b915051f7799a8520fc50f755ae42139e26ed7cd7e74df153d3792e4f8681e6d: Status 404 returned error can't find the container with id b915051f7799a8520fc50f755ae42139e26ed7cd7e74df153d3792e4f8681e6d Oct 04 07:10:05 crc kubenswrapper[4987]: W1004 07:10:05.376890 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-2135c82e5960a449112a49b7d694a5c54a3026020788c0619899fb07bcf0f1f9 WatchSource:0}: Error finding container 2135c82e5960a449112a49b7d694a5c54a3026020788c0619899fb07bcf0f1f9: Status 404 returned error can't find the container with id 2135c82e5960a449112a49b7d694a5c54a3026020788c0619899fb07bcf0f1f9 Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.405718 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.405759 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.405769 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.405785 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.405796 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.508831 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.508872 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.508883 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.508899 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.508910 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.530460 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.530598 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:10:06.530579492 +0000 UTC m=+23.585478334 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.613369 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.613417 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.613431 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.613452 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.613470 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.631131 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.631167 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.631193 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.631213 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631298 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631329 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631345 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631355 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631383 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631406 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631419 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631386 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:06.631366043 +0000 UTC m=+23.686264885 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631445 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631474 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:06.631462055 +0000 UTC m=+23.686360957 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631595 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:06.631571218 +0000 UTC m=+23.686470060 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:05 crc kubenswrapper[4987]: E1004 07:10:05.631609 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:06.631602779 +0000 UTC m=+23.686501621 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.715223 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.715276 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.715289 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.715321 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.715376 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.817615 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.817668 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.817679 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.817697 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.817711 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.856983 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qt4jr"] Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.857291 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.859047 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.859107 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.859493 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.870138 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.878565 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.889272 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.896736 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.907849 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.916936 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.919549 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.919567 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.919575 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.919590 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.919599 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:05Z","lastTransitionTime":"2025-10-04T07:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.927773 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.932943 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3beef3ad-079d-4f0b-90ed-ff72194c09ca-hosts-file\") pod \"node-resolver-qt4jr\" (UID: \"3beef3ad-079d-4f0b-90ed-ff72194c09ca\") " pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.933030 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66tb7\" (UniqueName: \"kubernetes.io/projected/3beef3ad-079d-4f0b-90ed-ff72194c09ca-kube-api-access-66tb7\") pod \"node-resolver-qt4jr\" (UID: \"3beef3ad-079d-4f0b-90ed-ff72194c09ca\") " pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:05 crc kubenswrapper[4987]: I1004 07:10:05.939612 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.022037 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.022083 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.022094 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.022110 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.022124 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.034517 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3beef3ad-079d-4f0b-90ed-ff72194c09ca-hosts-file\") pod \"node-resolver-qt4jr\" (UID: \"3beef3ad-079d-4f0b-90ed-ff72194c09ca\") " pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.034575 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66tb7\" (UniqueName: \"kubernetes.io/projected/3beef3ad-079d-4f0b-90ed-ff72194c09ca-kube-api-access-66tb7\") pod \"node-resolver-qt4jr\" (UID: \"3beef3ad-079d-4f0b-90ed-ff72194c09ca\") " pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.034654 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3beef3ad-079d-4f0b-90ed-ff72194c09ca-hosts-file\") pod \"node-resolver-qt4jr\" (UID: \"3beef3ad-079d-4f0b-90ed-ff72194c09ca\") " pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.054154 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66tb7\" (UniqueName: \"kubernetes.io/projected/3beef3ad-079d-4f0b-90ed-ff72194c09ca-kube-api-access-66tb7\") pod \"node-resolver-qt4jr\" (UID: \"3beef3ad-079d-4f0b-90ed-ff72194c09ca\") " pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.066840 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.067526 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.068769 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.069384 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.070567 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.071407 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.072061 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.072986 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.073606 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.074478 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.074979 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.076048 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.076646 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.077244 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.078295 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.078794 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.079695 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.080112 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.080681 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.081675 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.082105 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.083019 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.083422 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.084408 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.084917 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.085583 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.086650 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.087090 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.088070 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.088496 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.089297 4987 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.089419 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.090971 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.091964 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.092344 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.093900 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.094637 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.095607 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.096390 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.097391 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.097871 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.098791 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.099351 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.100314 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.100884 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.101996 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.102597 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.104121 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.104711 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.105914 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.106383 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.107312 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.108008 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.108477 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.125127 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.125171 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.125185 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.125199 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.125209 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.164749 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.164792 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"2135c82e5960a449112a49b7d694a5c54a3026020788c0619899fb07bcf0f1f9"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.166151 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b915051f7799a8520fc50f755ae42139e26ed7cd7e74df153d3792e4f8681e6d"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.167945 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qt4jr" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.168236 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.168294 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.168305 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ef3d8bac8e6d80b56ae09f5267982b0a56a9923e42f5b919a62ccd241c5f3fe1"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.170385 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.171174 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.172907 4987 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea" exitCode=255 Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.172976 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.173028 4987 scope.go:117] "RemoveContainer" containerID="410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.184877 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.185356 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.185803 4987 scope.go:117] "RemoveContainer" containerID="0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea" Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.185985 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.200531 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.214653 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-d76c6"] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.215511 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-dvv6j"] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.215743 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-mc8jj"] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.215965 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.216268 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.216315 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.218025 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.218191 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.218444 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.218548 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.218658 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.218951 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.219096 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.219211 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.219764 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.220327 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.220572 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.221278 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.221988 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.227415 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.227441 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.227450 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.227464 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.227543 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.237172 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.248349 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.263470 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.277021 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.291321 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.312774 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.329707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.329732 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.329740 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.329755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.329764 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337670 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-socket-dir-parent\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337715 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-kubelet\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337743 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-hostroot\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337762 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cni-binary-copy\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337788 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-cnibin\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337807 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-netns\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337885 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a89a1560-c018-4df7-92aa-795638d35d94-mcd-auth-proxy-config\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337933 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-cni-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337951 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-cni-bin\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337973 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgsnm\" (UniqueName: \"kubernetes.io/projected/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-kube-api-access-xgsnm\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.337992 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-k8s-cni-cncf-io\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338027 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-multus-certs\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338042 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phknb\" (UniqueName: \"kubernetes.io/projected/a89a1560-c018-4df7-92aa-795638d35d94-kube-api-access-phknb\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338059 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-cni-multus\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338093 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a89a1560-c018-4df7-92aa-795638d35d94-proxy-tls\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338112 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-etc-kubernetes\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338127 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cnibin\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338152 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338169 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-os-release\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338183 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/659797bb-8ec1-4ad8-880e-7aec3b068c46-cni-binary-copy\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338251 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338300 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-conf-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338317 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-system-cni-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338333 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n9dt\" (UniqueName: \"kubernetes.io/projected/659797bb-8ec1-4ad8-880e-7aec3b068c46-kube-api-access-7n9dt\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338363 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-daemon-config\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338378 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-system-cni-dir\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338392 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-os-release\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.338434 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a89a1560-c018-4df7-92aa-795638d35d94-rootfs\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.339549 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.363455 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.387306 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.410891 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.423792 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.431832 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.431870 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.431881 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.431898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.431909 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.436485 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438828 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-daemon-config\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438866 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-system-cni-dir\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438884 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-os-release\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438901 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a89a1560-c018-4df7-92aa-795638d35d94-rootfs\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438919 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-cnibin\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438933 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-socket-dir-parent\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438950 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-kubelet\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438964 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-hostroot\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438976 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-system-cni-dir\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.438980 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cni-binary-copy\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439036 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-netns\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439033 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-os-release\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439054 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-cni-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439074 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-cni-bin\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439072 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-kubelet\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439091 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a89a1560-c018-4df7-92aa-795638d35d94-mcd-auth-proxy-config\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439104 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-hostroot\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439135 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-cnibin\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439123 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgsnm\" (UniqueName: \"kubernetes.io/projected/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-kube-api-access-xgsnm\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439210 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-cni-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439078 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-netns\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439032 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a89a1560-c018-4df7-92aa-795638d35d94-rootfs\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439144 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-cni-bin\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439249 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-k8s-cni-cncf-io\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439277 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-multus-certs\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439304 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phknb\" (UniqueName: \"kubernetes.io/projected/a89a1560-c018-4df7-92aa-795638d35d94-kube-api-access-phknb\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439305 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-k8s-cni-cncf-io\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439328 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-run-multus-certs\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439330 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-cni-multus\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439355 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-host-var-lib-cni-multus\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439398 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a89a1560-c018-4df7-92aa-795638d35d94-proxy-tls\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439427 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-etc-kubernetes\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439446 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cnibin\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439464 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-os-release\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439479 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/659797bb-8ec1-4ad8-880e-7aec3b068c46-cni-binary-copy\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439497 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439517 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-conf-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439536 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439555 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-system-cni-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439557 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-etc-kubernetes\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439570 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n9dt\" (UniqueName: \"kubernetes.io/projected/659797bb-8ec1-4ad8-880e-7aec3b068c46-kube-api-access-7n9dt\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439655 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-conf-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439691 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a89a1560-c018-4df7-92aa-795638d35d94-mcd-auth-proxy-config\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439724 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-os-release\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439734 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cni-binary-copy\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439534 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cnibin\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.439810 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-system-cni-dir\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.440081 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-socket-dir-parent\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.440312 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.440342 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/659797bb-8ec1-4ad8-880e-7aec3b068c46-multus-daemon-config\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.440450 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.440467 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/659797bb-8ec1-4ad8-880e-7aec3b068c46-cni-binary-copy\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.443677 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a89a1560-c018-4df7-92aa-795638d35d94-proxy-tls\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.456741 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phknb\" (UniqueName: \"kubernetes.io/projected/a89a1560-c018-4df7-92aa-795638d35d94-kube-api-access-phknb\") pod \"machine-config-daemon-dvv6j\" (UID: \"a89a1560-c018-4df7-92aa-795638d35d94\") " pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.456859 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgsnm\" (UniqueName: \"kubernetes.io/projected/9fad7c88-7e3a-4671-8e85-b6e228b2c9c6-kube-api-access-xgsnm\") pod \"multus-additional-cni-plugins-d76c6\" (UID: \"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\") " pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.456860 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n9dt\" (UniqueName: \"kubernetes.io/projected/659797bb-8ec1-4ad8-880e-7aec3b068c46-kube-api-access-7n9dt\") pod \"multus-mc8jj\" (UID: \"659797bb-8ec1-4ad8-880e-7aec3b068c46\") " pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.460333 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.475694 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.491080 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:09:58Z\\\",\\\"message\\\":\\\"W1004 07:09:47.153085 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 07:09:47.153566 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759561787 cert, and key in /tmp/serving-cert-983044263/serving-signer.crt, /tmp/serving-cert-983044263/serving-signer.key\\\\nI1004 07:09:47.560049 1 observer_polling.go:159] Starting file observer\\\\nW1004 07:09:47.569062 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 07:09:47.569261 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:09:47.571583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-983044263/tls.crt::/tmp/serving-cert-983044263/tls.key\\\\\\\"\\\\nF1004 07:09:57.798536 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.505897 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.518794 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.528983 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-mc8jj" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.534703 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.535049 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.535064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.535086 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.535100 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.535931 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-d76c6" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.540632 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.540754 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:10:08.540730743 +0000 UTC m=+25.595629585 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.545046 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.591340 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tw9rc"] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.592551 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.594773 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.595017 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.595200 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.595545 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.595758 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.596318 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.598608 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.606379 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.616965 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.631221 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.638075 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.638135 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.638149 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.638174 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.638195 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641456 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641508 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641685 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-systemd\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641715 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-log-socket\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.641713 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.641773 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.641789 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641809 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovn-node-metrics-cert\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641842 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-var-lib-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641868 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-env-overrides\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641899 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-kubelet\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.641927 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-config\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642008 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8rdv\" (UniqueName: \"kubernetes.io/projected/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-kube-api-access-k8rdv\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642121 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642165 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642199 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-script-lib\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642226 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642278 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:08.642257315 +0000 UTC m=+25.697156157 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642297 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642359 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642424 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:08.64240769 +0000 UTC m=+25.697306532 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642470 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642482 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:08.642469291 +0000 UTC m=+25.697368133 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642488 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642501 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:06 crc kubenswrapper[4987]: E1004 07:10:06.642556 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:08.642541094 +0000 UTC m=+25.697439936 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642659 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-ovn\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642754 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-netd\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642840 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-systemd-units\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642869 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-bin\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642906 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-etc-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642932 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-slash\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642958 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-node-log\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.642991 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-netns\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.643018 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.643049 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-ovn-kubernetes\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.650193 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: W1004 07:10:06.662236 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659797bb_8ec1_4ad8_880e_7aec3b068c46.slice/crio-d8cdf214bd3bb262b49847034c7ff2f3cc7bd07802d18a2696b9dd91c4336b8d WatchSource:0}: Error finding container d8cdf214bd3bb262b49847034c7ff2f3cc7bd07802d18a2696b9dd91c4336b8d: Status 404 returned error can't find the container with id d8cdf214bd3bb262b49847034c7ff2f3cc7bd07802d18a2696b9dd91c4336b8d Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.664367 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.682604 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.698814 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.711081 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.726881 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:09:58Z\\\",\\\"message\\\":\\\"W1004 07:09:47.153085 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 07:09:47.153566 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759561787 cert, and key in /tmp/serving-cert-983044263/serving-signer.crt, /tmp/serving-cert-983044263/serving-signer.key\\\\nI1004 07:09:47.560049 1 observer_polling.go:159] Starting file observer\\\\nW1004 07:09:47.569062 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 07:09:47.569261 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:09:47.571583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-983044263/tls.crt::/tmp/serving-cert-983044263/tls.key\\\\\\\"\\\\nF1004 07:09:57.798536 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.741016 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743563 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743602 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743613 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743664 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743732 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-ovn-kubernetes\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743775 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743812 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-systemd\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743833 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-log-socket\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743854 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovn-node-metrics-cert\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743875 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-var-lib-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743895 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-env-overrides\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743927 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-kubelet\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743917 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-ovn-kubernetes\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743963 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-log-socket\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743950 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-config\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743980 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743931 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-systemd\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744018 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8rdv\" (UniqueName: \"kubernetes.io/projected/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-kube-api-access-k8rdv\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744210 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-kubelet\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744283 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-script-lib\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744335 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-ovn\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744350 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-var-lib-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744398 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-netd\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744409 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-ovn\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744368 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-netd\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744485 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-systemd-units\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744510 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-bin\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744551 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-etc-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744579 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-slash\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744587 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-systemd-units\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744601 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-node-log\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744602 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-bin\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744657 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-netns\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744677 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-slash\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744687 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744713 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-etc-openvswitch\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744759 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-netns\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744766 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-node-log\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744799 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.743678 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.744965 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-env-overrides\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.745027 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-config\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.745349 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-script-lib\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.747998 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovn-node-metrics-cert\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.757386 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.768323 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8rdv\" (UniqueName: \"kubernetes.io/projected/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-kube-api-access-k8rdv\") pod \"ovnkube-node-tw9rc\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.772731 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.795039 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.847467 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.847501 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.847508 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.847522 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.847532 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.904480 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:06 crc kubenswrapper[4987]: W1004 07:10:06.919673 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8b42ffd_33d6_48ad_9a54_1bfad0c6bba3.slice/crio-f1084503219d9dab387d1525440270a1effad654748dbe8c4b08caa6def47c1f WatchSource:0}: Error finding container f1084503219d9dab387d1525440270a1effad654748dbe8c4b08caa6def47c1f: Status 404 returned error can't find the container with id f1084503219d9dab387d1525440270a1effad654748dbe8c4b08caa6def47c1f Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.950644 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.950693 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.950706 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.950730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:06 crc kubenswrapper[4987]: I1004 07:10:06.950745 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:06Z","lastTransitionTime":"2025-10-04T07:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.052440 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.052474 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:07 crc kubenswrapper[4987]: E1004 07:10:07.052555 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.052899 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:07 crc kubenswrapper[4987]: E1004 07:10:07.052951 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:07 crc kubenswrapper[4987]: E1004 07:10:07.053067 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.053800 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.053841 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.053852 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.053874 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.053885 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.156064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.156109 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.156119 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.156166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.156177 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.177331 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.177390 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.177556 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"4e7813e43433783aa5a0090490ab102c056b088f4d52a1cd368403dba9df827e"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.179067 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qt4jr" event={"ID":"3beef3ad-079d-4f0b-90ed-ff72194c09ca","Type":"ContainerStarted","Data":"33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.179095 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qt4jr" event={"ID":"3beef3ad-079d-4f0b-90ed-ff72194c09ca","Type":"ContainerStarted","Data":"75ad0f480dd60a195ede7905722e8507a0394e6bfe4926c8c4ce1e7b486aa098"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.180595 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.183237 4987 scope.go:117] "RemoveContainer" containerID="0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea" Oct 04 07:10:07 crc kubenswrapper[4987]: E1004 07:10:07.183902 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.193250 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerStarted","Data":"d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.193311 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerStarted","Data":"d8cdf214bd3bb262b49847034c7ff2f3cc7bd07802d18a2696b9dd91c4336b8d"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.195443 4987 generic.go:334] "Generic (PLEG): container finished" podID="9fad7c88-7e3a-4671-8e85-b6e228b2c9c6" containerID="233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a" exitCode=0 Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.195535 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerDied","Data":"233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.195571 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerStarted","Data":"e4fc9ae67d4a20a94586bca93d2f27d658df8541af857f66adbdf9551c515f62"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.198143 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" exitCode=0 Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.198185 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.198211 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"f1084503219d9dab387d1525440270a1effad654748dbe8c4b08caa6def47c1f"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.208425 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.222469 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.238073 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.255064 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.258547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.258575 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.258584 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.258600 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.258609 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.272549 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://410dda75710575344c92312b1cd81a9b807fe8f80439f54e1dcccfe852661845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:09:58Z\\\",\\\"message\\\":\\\"W1004 07:09:47.153085 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1004 07:09:47.153566 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759561787 cert, and key in /tmp/serving-cert-983044263/serving-signer.crt, /tmp/serving-cert-983044263/serving-signer.key\\\\nI1004 07:09:47.560049 1 observer_polling.go:159] Starting file observer\\\\nW1004 07:09:47.569062 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1004 07:09:47.569261 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:09:47.571583 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-983044263/tls.crt::/tmp/serving-cert-983044263/tls.key\\\\\\\"\\\\nF1004 07:09:57.798536 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.285604 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.299850 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.322296 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.351281 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.361335 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.361385 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.361396 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.361413 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.361424 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.369564 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.389732 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.409911 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.424785 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.439546 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.459509 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.463994 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.464024 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.464034 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.464050 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.464061 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.482909 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.507172 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.519915 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.528022 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.532291 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.543141 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.559063 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.566815 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.567205 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.567216 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.567233 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.567242 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.573588 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.592922 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.608314 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.623897 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.639071 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:07Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.670140 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.670186 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.670195 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.670210 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.670221 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.774928 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.774985 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.774999 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.775021 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.775035 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.878972 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.879030 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.879047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.879067 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.879079 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.980955 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.981008 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.981020 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.981039 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:07 crc kubenswrapper[4987]: I1004 07:10:07.981049 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:07Z","lastTransitionTime":"2025-10-04T07:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.083796 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.084161 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.084171 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.084189 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.084201 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.185955 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.186011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.186024 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.186040 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.186052 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.203925 4987 generic.go:334] "Generic (PLEG): container finished" podID="9fad7c88-7e3a-4671-8e85-b6e228b2c9c6" containerID="e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62" exitCode=0 Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.203998 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerDied","Data":"e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.205709 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.210636 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.210684 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.210702 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.210713 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.210723 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.210734 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.211267 4987 scope.go:117] "RemoveContainer" containerID="0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea" Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.211479 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.219570 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.235238 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.249251 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.271030 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.285643 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.290510 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.290538 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.290548 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.290563 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.290575 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.301662 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.311542 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.327201 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.342333 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.357790 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.370353 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.383329 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.393077 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.393118 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.393128 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.393143 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.393152 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.395850 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.407606 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.420208 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.430358 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.441462 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.450386 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.465457 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.480439 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.495477 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.495505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.495514 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.495527 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.495536 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.497107 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.519165 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.522032 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-fx5cm"] Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.522438 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.527021 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.527124 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.527282 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.527482 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.534818 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.547389 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.559036 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.562937 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.563072 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6frph\" (UniqueName: \"kubernetes.io/projected/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-kube-api-access-6frph\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.563086 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:10:12.563061275 +0000 UTC m=+29.617960127 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.563112 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-serviceca\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.563204 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-host\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.574139 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.590255 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.599194 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.599384 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.599446 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.599465 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.599474 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.604528 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.634315 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.649123 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.662078 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664331 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664371 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664391 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664412 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664428 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-host\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664454 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6frph\" (UniqueName: \"kubernetes.io/projected/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-kube-api-access-6frph\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.664477 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-serviceca\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.664729 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.664803 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:12.664786023 +0000 UTC m=+29.719684865 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665059 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665085 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665096 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665128 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:12.665119023 +0000 UTC m=+29.720017965 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665177 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665205 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:12.665197525 +0000 UTC m=+29.720096467 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.665251 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-host\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665330 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665351 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665361 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:08 crc kubenswrapper[4987]: E1004 07:10:08.665390 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:12.66538097 +0000 UTC m=+29.720279902 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.665867 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-serviceca\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.674305 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.686300 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6frph\" (UniqueName: \"kubernetes.io/projected/f1bc8f96-f083-44b7-8b6d-6954d327dc2c-kube-api-access-6frph\") pod \"node-ca-fx5cm\" (UID: \"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\") " pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.687944 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.702100 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.702580 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.702660 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.702671 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.702686 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.702695 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.719266 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.733457 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.747448 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.760152 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.771259 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.805937 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.805972 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.805981 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.805998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.806008 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.807104 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:08Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.841268 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-fx5cm" Oct 04 07:10:08 crc kubenswrapper[4987]: W1004 07:10:08.857249 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1bc8f96_f083_44b7_8b6d_6954d327dc2c.slice/crio-c067d049b25de5d054a4fa350a3c433a385304ba8bd74d0bfa3793791f7ca7e6 WatchSource:0}: Error finding container c067d049b25de5d054a4fa350a3c433a385304ba8bd74d0bfa3793791f7ca7e6: Status 404 returned error can't find the container with id c067d049b25de5d054a4fa350a3c433a385304ba8bd74d0bfa3793791f7ca7e6 Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.908653 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.908690 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.908699 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.908719 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:08 crc kubenswrapper[4987]: I1004 07:10:08.908728 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:08Z","lastTransitionTime":"2025-10-04T07:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.011388 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.011879 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.011893 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.011910 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.011923 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.053025 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.053159 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.053268 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:09 crc kubenswrapper[4987]: E1004 07:10:09.053255 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:09 crc kubenswrapper[4987]: E1004 07:10:09.053373 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:09 crc kubenswrapper[4987]: E1004 07:10:09.053464 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.114877 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.114916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.114932 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.114947 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.114957 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.214783 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fx5cm" event={"ID":"f1bc8f96-f083-44b7-8b6d-6954d327dc2c","Type":"ContainerStarted","Data":"59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.214835 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-fx5cm" event={"ID":"f1bc8f96-f083-44b7-8b6d-6954d327dc2c","Type":"ContainerStarted","Data":"c067d049b25de5d054a4fa350a3c433a385304ba8bd74d0bfa3793791f7ca7e6"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.216763 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.216800 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.216808 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.216822 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.216831 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.218360 4987 generic.go:334] "Generic (PLEG): container finished" podID="9fad7c88-7e3a-4671-8e85-b6e228b2c9c6" containerID="35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96" exitCode=0 Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.218416 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerDied","Data":"35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.234390 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.248767 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.264375 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.280570 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.293490 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.303319 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.315030 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.318998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.319042 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.319054 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.319069 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.319080 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.327958 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.342546 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.363445 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.389932 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.405012 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.421679 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.421718 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.421727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.421742 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.421751 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.423750 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.440564 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.452100 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.463742 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.486560 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.523534 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.523601 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.523612 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.523642 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.523656 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.529751 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.568050 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.610770 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.626018 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.626059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.626072 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.626089 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.626100 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.647609 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.687678 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.728024 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.728070 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.728081 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.728099 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.728111 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.729774 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.780961 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.811009 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.831639 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.831680 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.831691 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.831708 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.831719 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.847022 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.886452 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.931154 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:09Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.933366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.933393 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.933401 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.933422 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:09 crc kubenswrapper[4987]: I1004 07:10:09.933434 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:09Z","lastTransitionTime":"2025-10-04T07:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.036543 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.036580 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.036594 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.036613 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.036643 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.139558 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.139604 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.139616 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.139675 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.139690 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.224284 4987 generic.go:334] "Generic (PLEG): container finished" podID="9fad7c88-7e3a-4671-8e85-b6e228b2c9c6" containerID="ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b" exitCode=0 Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.224358 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerDied","Data":"ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.231583 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.241869 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.242017 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.242080 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.242142 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.242196 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.247274 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.263061 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.286153 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.299236 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.311935 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.326036 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.336053 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.344804 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.344837 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.344849 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.344865 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.344876 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.349036 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.361347 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.371867 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.384139 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.407725 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.448485 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.448522 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.448531 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.448547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.448558 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.449390 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.488721 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:10Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.550741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.551064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.551077 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.551096 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.551108 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.653655 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.653698 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.653708 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.653724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.653734 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.756543 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.756587 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.756597 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.756615 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.756644 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.859696 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.859745 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.859762 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.859780 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.859793 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.962162 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.962221 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.962235 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.962253 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:10 crc kubenswrapper[4987]: I1004 07:10:10.962292 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:10Z","lastTransitionTime":"2025-10-04T07:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.053242 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.053252 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:11 crc kubenswrapper[4987]: E1004 07:10:11.053389 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.053420 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:11 crc kubenswrapper[4987]: E1004 07:10:11.053510 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:11 crc kubenswrapper[4987]: E1004 07:10:11.053598 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.064287 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.064338 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.064351 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.064370 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.064386 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.167952 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.168012 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.168033 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.168059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.168076 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.237173 4987 generic.go:334] "Generic (PLEG): container finished" podID="9fad7c88-7e3a-4671-8e85-b6e228b2c9c6" containerID="81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046" exitCode=0 Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.237231 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerDied","Data":"81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.252892 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.271728 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.271768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.271779 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.271794 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.271804 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.274120 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.287244 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.303111 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.314382 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.338503 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.353990 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.369900 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.373868 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.373899 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.373909 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.373924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.373934 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.381326 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.394202 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.404440 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.414442 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.425546 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.436106 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:11Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.476217 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.476260 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.476271 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.476289 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.476300 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.579417 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.580014 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.580098 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.580179 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.580252 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.682541 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.682585 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.682595 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.682611 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.682641 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.784855 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.784890 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.784899 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.784912 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.784921 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.887535 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.887578 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.887597 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.887614 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.887640 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.990094 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.990153 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.990167 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.990188 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:11 crc kubenswrapper[4987]: I1004 07:10:11.990204 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:11Z","lastTransitionTime":"2025-10-04T07:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.093585 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.093704 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.093723 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.093754 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.093779 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.196651 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.196909 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.197028 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.197140 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.197239 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.245565 4987 generic.go:334] "Generic (PLEG): container finished" podID="9fad7c88-7e3a-4671-8e85-b6e228b2c9c6" containerID="f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494" exitCode=0 Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.245665 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerDied","Data":"f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.263752 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.286380 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.299700 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.299741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.299754 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.299774 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.299786 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.303057 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.317742 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.331434 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.345953 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.358845 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.370556 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.386215 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.405311 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.405355 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.405365 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.405382 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.405394 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.406211 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.422936 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.437366 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.451755 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.466871 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:12Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.507383 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.507430 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.507440 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.507459 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.507470 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.605583 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.605811 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:10:20.605792403 +0000 UTC m=+37.660691245 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.610563 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.610615 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.610657 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.610677 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.610691 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.706908 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.706980 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.707018 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.707055 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707097 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707133 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707152 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707159 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707185 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707216 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707231 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707246 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707232 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:20.707209702 +0000 UTC m=+37.762108564 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707429 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:20.707362946 +0000 UTC m=+37.762261798 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707485 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:20.707467749 +0000 UTC m=+37.762366601 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:12 crc kubenswrapper[4987]: E1004 07:10:12.707504 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:20.70749465 +0000 UTC m=+37.762393502 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.714049 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.714112 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.714166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.714204 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.714231 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.816838 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.816879 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.816888 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.816905 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.816914 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.919704 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.919788 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.919801 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.919824 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:12 crc kubenswrapper[4987]: I1004 07:10:12.919840 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:12Z","lastTransitionTime":"2025-10-04T07:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.023166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.023209 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.023218 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.023233 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.023242 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.052760 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.052782 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.052893 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:13 crc kubenswrapper[4987]: E1004 07:10:13.053024 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:13 crc kubenswrapper[4987]: E1004 07:10:13.053165 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:13 crc kubenswrapper[4987]: E1004 07:10:13.053296 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.125638 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.125671 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.125681 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.125696 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.125726 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.228446 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.228483 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.228492 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.228504 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.228512 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.256111 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" event={"ID":"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6","Type":"ContainerStarted","Data":"2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.260723 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.261009 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.261058 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.269473 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.279416 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.285403 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.293367 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.308709 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.322977 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.331818 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.331868 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.331882 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.331916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.331930 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.335450 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.348857 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.363846 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.379250 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.399716 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.413153 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.431283 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.434507 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.434558 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.434569 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.434611 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.434635 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.449929 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.465915 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.481070 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.495473 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.508249 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.523683 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.537418 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.537702 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.537798 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.537878 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.537964 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.539490 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.552103 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.567418 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.579551 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.598350 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.611762 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.629496 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.640331 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.640583 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.640710 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.640805 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.640874 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.641184 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.653307 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.669046 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:13Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.743481 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.743693 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.743760 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.743821 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.743884 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.847430 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.847931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.848143 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.848362 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.848517 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.952031 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.952090 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.952102 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.952119 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:13 crc kubenswrapper[4987]: I1004 07:10:13.952131 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:13Z","lastTransitionTime":"2025-10-04T07:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.054575 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.054926 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.055042 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.055190 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.055310 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.073468 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.100048 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.114433 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.132166 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.147470 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.157347 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.157420 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.157439 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.157468 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.157488 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.168728 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.191354 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.210532 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.229372 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.248734 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.259648 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.259686 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.259698 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.259715 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.259726 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.263161 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.264006 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.278265 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.289416 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.292964 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.304335 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.319620 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.333565 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.344588 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.361425 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.361463 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.361478 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.361497 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.361509 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.363397 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.378542 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.392292 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.406681 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.419599 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.433726 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.444169 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.458101 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.463817 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.463867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.463885 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.463954 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.463974 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.474478 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.488719 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.512360 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.566025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.566068 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.566078 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.566093 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.566106 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.668493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.668595 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.668612 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.668650 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.668671 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.771226 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.771266 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.771277 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.771293 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.771304 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.873845 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.873899 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.873915 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.873934 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.873946 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.907433 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.907478 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.907490 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.907506 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.907517 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: E1004 07:10:14.921703 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.925663 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.925702 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.925714 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.925750 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.925765 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: E1004 07:10:14.942259 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.945328 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.945360 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.945369 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.945387 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.945396 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: E1004 07:10:14.956842 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.960280 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.960321 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.960335 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.960351 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.960361 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: E1004 07:10:14.977819 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.981656 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.981723 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.981737 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.981753 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.981764 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:14 crc kubenswrapper[4987]: E1004 07:10:14.993336 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:14 crc kubenswrapper[4987]: E1004 07:10:14.993490 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.995058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.995110 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.995121 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.995136 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:14 crc kubenswrapper[4987]: I1004 07:10:14.995147 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:14Z","lastTransitionTime":"2025-10-04T07:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.052855 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.052893 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:15 crc kubenswrapper[4987]: E1004 07:10:15.052973 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.052855 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:15 crc kubenswrapper[4987]: E1004 07:10:15.053052 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:15 crc kubenswrapper[4987]: E1004 07:10:15.053109 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.097371 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.097416 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.097430 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.097449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.097461 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.199870 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.199906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.199916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.199930 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.199939 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.302488 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.302541 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.302554 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.302573 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.302586 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.405021 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.405050 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.405059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.405072 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.405081 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.507998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.508223 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.508236 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.508248 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.508258 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.610366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.610402 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.610411 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.610424 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.610433 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.712353 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.712415 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.712424 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.712438 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.712447 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.814792 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.814829 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.814838 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.814855 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.814864 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.917663 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.917730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.917743 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.917783 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:15 crc kubenswrapper[4987]: I1004 07:10:15.917795 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:15Z","lastTransitionTime":"2025-10-04T07:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.020660 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.020698 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.020710 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.020733 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.020745 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.123920 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.123958 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.123967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.123982 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.123992 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.226377 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.226413 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.226424 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.226438 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.226448 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.270305 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/0.log" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.272823 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5" exitCode=1 Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.272878 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.274186 4987 scope.go:117] "RemoveContainer" containerID="e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.295750 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.315228 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.328218 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.328269 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.328286 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.328308 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.328323 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.332298 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.351775 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:16Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 07:10:15.778125 6260 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 07:10:15.778251 6260 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 07:10:15.778541 6260 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 07:10:15.778555 6260 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 07:10:15.778572 6260 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 07:10:15.778582 6260 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 07:10:15.778586 6260 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 07:10:15.778605 6260 factory.go:656] Stopping watch factory\\\\nI1004 07:10:15.778616 6260 ovnkube.go:599] Stopped ovnkube\\\\nI1004 07:10:15.778611 6260 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 07:10:15.778639 6260 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 07:10:15.778653 6260 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 07:10:15.778659 6260 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.363653 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.378207 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.389806 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.409399 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.430600 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.430840 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.430876 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.430887 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.430903 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.430914 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.447130 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.457010 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.471102 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.485535 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.497850 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:16Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.533590 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.533657 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.533671 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.533694 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.533705 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.636015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.636073 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.636084 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.636103 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.636438 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.739211 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.739239 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.739249 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.739263 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.739273 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.841334 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.841387 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.841399 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.841419 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.841430 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.943669 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.943703 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.943712 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.943725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:16 crc kubenswrapper[4987]: I1004 07:10:16.943734 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:16Z","lastTransitionTime":"2025-10-04T07:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.046477 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.046514 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.046528 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.046545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.046555 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.052845 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.052894 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:17 crc kubenswrapper[4987]: E1004 07:10:17.052955 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.053024 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:17 crc kubenswrapper[4987]: E1004 07:10:17.053109 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:17 crc kubenswrapper[4987]: E1004 07:10:17.053176 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.149088 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.149134 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.149149 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.149166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.149179 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.251860 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.251902 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.251911 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.251927 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.251943 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.277141 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/1.log" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.277796 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/0.log" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.280690 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2" exitCode=1 Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.280736 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.280776 4987 scope.go:117] "RemoveContainer" containerID="e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.281243 4987 scope.go:117] "RemoveContainer" containerID="eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2" Oct 04 07:10:17 crc kubenswrapper[4987]: E1004 07:10:17.281506 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.295746 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.310416 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.319191 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.332998 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.350417 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.354852 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.354892 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.354903 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.354922 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.354933 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.367391 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.380834 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.394127 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.406075 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.419030 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.435455 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.452518 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.458215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.458259 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.458272 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.458296 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.458315 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.468838 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.488449 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e90b95cf847bb6b037a625c6da57862e40b4b8858c6803970f63062c940123c5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:16Z\\\",\\\"message\\\":\\\"lector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 07:10:15.778125 6260 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 07:10:15.778251 6260 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 07:10:15.778541 6260 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 07:10:15.778555 6260 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 07:10:15.778572 6260 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 07:10:15.778582 6260 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 07:10:15.778586 6260 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 07:10:15.778605 6260 factory.go:656] Stopping watch factory\\\\nI1004 07:10:15.778616 6260 ovnkube.go:599] Stopped ovnkube\\\\nI1004 07:10:15.778611 6260 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 07:10:15.778639 6260 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 07:10:15.778653 6260 handler.go:208] Removed *v1.Node event handler 7\\\\nI1004 07:10:15.778659 6260 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:17Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.562126 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.562166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.562177 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.562196 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.562206 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.664944 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.664976 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.664984 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.664999 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.665009 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.767065 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.767107 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.767118 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.767136 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.767147 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.869046 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.869084 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.869095 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.869111 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.869121 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.971107 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.971165 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.971183 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.971207 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:17 crc kubenswrapper[4987]: I1004 07:10:17.971224 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:17Z","lastTransitionTime":"2025-10-04T07:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.074352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.074393 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.074404 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.074423 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.074433 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.177216 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.177257 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.177267 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.177286 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.177299 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.280296 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.280351 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.280361 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.280376 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.280386 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.285669 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/1.log" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.289145 4987 scope.go:117] "RemoveContainer" containerID="eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2" Oct 04 07:10:18 crc kubenswrapper[4987]: E1004 07:10:18.289302 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.302228 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.317098 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.331096 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.344650 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.356894 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.373474 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.384675 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.384724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.384740 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.384759 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.384778 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.389243 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.403158 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.416007 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.436582 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.450714 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.463972 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.475352 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.486727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.486761 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.486772 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.486788 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.486799 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.490054 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.589390 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.589464 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.589478 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.589495 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.589508 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.597508 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm"] Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.598052 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.600136 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.600838 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.619558 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.639300 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.653324 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.666848 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8738c81b-0365-44fd-bd47-1a1a3b1716d0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.666951 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8738c81b-0365-44fd-bd47-1a1a3b1716d0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.667010 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzlrr\" (UniqueName: \"kubernetes.io/projected/8738c81b-0365-44fd-bd47-1a1a3b1716d0-kube-api-access-tzlrr\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.667037 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8738c81b-0365-44fd-bd47-1a1a3b1716d0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.668873 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.682650 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.691936 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.691989 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.692005 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.692025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.692040 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.697357 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.710708 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.721359 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.734488 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.747437 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.758557 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.767848 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8738c81b-0365-44fd-bd47-1a1a3b1716d0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.767899 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8738c81b-0365-44fd-bd47-1a1a3b1716d0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.767945 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzlrr\" (UniqueName: \"kubernetes.io/projected/8738c81b-0365-44fd-bd47-1a1a3b1716d0-kube-api-access-tzlrr\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.767970 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8738c81b-0365-44fd-bd47-1a1a3b1716d0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.768635 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8738c81b-0365-44fd-bd47-1a1a3b1716d0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.768700 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8738c81b-0365-44fd-bd47-1a1a3b1716d0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.771764 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.773489 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8738c81b-0365-44fd-bd47-1a1a3b1716d0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.782762 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzlrr\" (UniqueName: \"kubernetes.io/projected/8738c81b-0365-44fd-bd47-1a1a3b1716d0-kube-api-access-tzlrr\") pod \"ovnkube-control-plane-749d76644c-k48xm\" (UID: \"8738c81b-0365-44fd-bd47-1a1a3b1716d0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.784104 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.795147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.795193 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.795203 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.795221 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.795232 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.796881 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.813913 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:18Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.898514 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.898584 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.898598 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.898654 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.898666 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:18Z","lastTransitionTime":"2025-10-04T07:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:18 crc kubenswrapper[4987]: I1004 07:10:18.909951 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" Oct 04 07:10:18 crc kubenswrapper[4987]: W1004 07:10:18.923347 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8738c81b_0365_44fd_bd47_1a1a3b1716d0.slice/crio-e2a5a7c5e94bae581e3df6f76c03c4a407b7067acdd26152fa16eea5100ff344 WatchSource:0}: Error finding container e2a5a7c5e94bae581e3df6f76c03c4a407b7067acdd26152fa16eea5100ff344: Status 404 returned error can't find the container with id e2a5a7c5e94bae581e3df6f76c03c4a407b7067acdd26152fa16eea5100ff344 Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.002894 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.002926 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.002952 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.002966 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.002975 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.053570 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.053667 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.053648 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:19 crc kubenswrapper[4987]: E1004 07:10:19.053830 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:19 crc kubenswrapper[4987]: E1004 07:10:19.053950 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:19 crc kubenswrapper[4987]: E1004 07:10:19.054034 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.105513 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.105548 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.105557 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.105572 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.105584 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.208274 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.208341 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.208351 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.208366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.208383 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.292643 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" event={"ID":"8738c81b-0365-44fd-bd47-1a1a3b1716d0","Type":"ContainerStarted","Data":"e2a5a7c5e94bae581e3df6f76c03c4a407b7067acdd26152fa16eea5100ff344"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.312141 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.312205 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.312222 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.312244 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.312257 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.414602 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.414667 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.414682 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.414706 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.414720 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.516942 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.516988 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.517002 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.517020 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.517033 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.619547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.619601 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.619611 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.619647 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.619659 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.721804 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.721853 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.721867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.721886 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.721898 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.823715 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.823766 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.823775 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.823791 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.823804 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.926298 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.926337 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.926349 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.926367 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:19 crc kubenswrapper[4987]: I1004 07:10:19.926378 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:19Z","lastTransitionTime":"2025-10-04T07:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.028703 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.028748 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.028763 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.028781 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.028792 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.131087 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.131120 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.131128 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.131329 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.131339 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.233324 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.233383 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.233394 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.233412 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.233423 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.297140 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" event={"ID":"8738c81b-0365-44fd-bd47-1a1a3b1716d0","Type":"ContainerStarted","Data":"cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.297181 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" event={"ID":"8738c81b-0365-44fd-bd47-1a1a3b1716d0","Type":"ContainerStarted","Data":"6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.308906 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.324414 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.335867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.335935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.335953 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.335977 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.335998 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.337321 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.348677 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.359822 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.368935 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.379523 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.389872 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.399190 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.400792 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-m6jjp"] Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.401235 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.401299 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.418429 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.428297 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.438186 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.438231 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.438242 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.438258 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.438267 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.438749 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.452798 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.464117 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.484155 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.486838 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t456h\" (UniqueName: \"kubernetes.io/projected/4b4d7a8a-0a67-4831-9a78-1369771b81db-kube-api-access-t456h\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.486885 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.496056 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.512207 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.524615 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.540806 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.540852 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.540868 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.540892 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.540910 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.543055 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.555604 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.568494 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.579022 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.587657 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t456h\" (UniqueName: \"kubernetes.io/projected/4b4d7a8a-0a67-4831-9a78-1369771b81db-kube-api-access-t456h\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.587733 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.588011 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.588102 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:10:21.088081327 +0000 UTC m=+38.142980189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.598877 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.611911 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.614070 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t456h\" (UniqueName: \"kubernetes.io/projected/4b4d7a8a-0a67-4831-9a78-1369771b81db-kube-api-access-t456h\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.624271 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.634426 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.643129 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.643157 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.643166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.643180 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.643192 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.645718 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.658688 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.668661 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.677847 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.688783 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.688967 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:10:36.68894322 +0000 UTC m=+53.743842092 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.690511 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:20Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.745852 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.745901 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.745909 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.745924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.745934 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.790563 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.790742 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.790811 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.790852 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.790903 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791004 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791031 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791049 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791052 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791007 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:36.790978238 +0000 UTC m=+53.845877120 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791154 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:36.791125252 +0000 UTC m=+53.846024144 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791189 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:36.791171883 +0000 UTC m=+53.846070765 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791504 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791536 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791551 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:20 crc kubenswrapper[4987]: E1004 07:10:20.791610 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:10:36.791589745 +0000 UTC m=+53.846488677 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.848606 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.848720 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.848751 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.848780 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.848801 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.951589 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.951714 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.951742 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.951776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:20 crc kubenswrapper[4987]: I1004 07:10:20.951802 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:20Z","lastTransitionTime":"2025-10-04T07:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.053279 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.053315 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.053518 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:21 crc kubenswrapper[4987]: E1004 07:10:21.053692 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.053810 4987 scope.go:117] "RemoveContainer" containerID="0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea" Oct 04 07:10:21 crc kubenswrapper[4987]: E1004 07:10:21.054086 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:21 crc kubenswrapper[4987]: E1004 07:10:21.054156 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.055360 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.055385 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.055396 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.055411 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.055424 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.095026 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:21 crc kubenswrapper[4987]: E1004 07:10:21.095309 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:21 crc kubenswrapper[4987]: E1004 07:10:21.095436 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:10:22.09540687 +0000 UTC m=+39.150305752 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.158272 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.158324 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.158337 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.158356 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.158369 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.261164 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.261213 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.261225 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.261243 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.261255 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.363738 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.363768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.363778 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.363795 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.363807 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.465579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.465709 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.465725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.465749 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.465771 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.568138 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.568174 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.568183 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.568197 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.568205 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.670489 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.670850 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.670862 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.670884 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.670896 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.772967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.773010 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.773022 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.773039 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.773051 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.876082 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.876126 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.876140 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.876160 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.876174 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.978675 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.978705 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.978714 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.978727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:21 crc kubenswrapper[4987]: I1004 07:10:21.978736 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:21Z","lastTransitionTime":"2025-10-04T07:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.053123 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:22 crc kubenswrapper[4987]: E1004 07:10:22.053454 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.081409 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.081445 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.081458 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.081473 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.081485 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.104970 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:22 crc kubenswrapper[4987]: E1004 07:10:22.105086 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:22 crc kubenswrapper[4987]: E1004 07:10:22.105125 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:10:24.105112768 +0000 UTC m=+41.160011610 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.184736 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.184806 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.184818 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.184841 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.184857 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.287922 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.287965 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.287974 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.287992 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.288001 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.303896 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.305458 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.305844 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.318508 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.333774 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.344099 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.356251 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.372748 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.387551 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.390011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.390045 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.390057 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.390076 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.390090 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.402016 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.445772 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.463997 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.485090 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.492461 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.492500 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.492511 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.492529 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.492540 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.502150 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.515209 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.529342 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.542723 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.556043 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.577268 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:22Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.595263 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.595345 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.595373 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.595404 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.595427 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.698488 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.698562 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.698587 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.698619 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.698678 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.801375 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.801417 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.801430 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.801449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.801462 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.904819 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.904891 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.904912 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.904938 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:22 crc kubenswrapper[4987]: I1004 07:10:22.904957 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:22Z","lastTransitionTime":"2025-10-04T07:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.007050 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.007086 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.007096 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.007112 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.007121 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.052405 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.052476 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:23 crc kubenswrapper[4987]: E1004 07:10:23.052534 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.052413 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:23 crc kubenswrapper[4987]: E1004 07:10:23.052601 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:23 crc kubenswrapper[4987]: E1004 07:10:23.052652 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.111025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.111064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.111074 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.111092 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.111106 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.213391 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.213604 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.213718 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.213815 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.213887 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.316730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.316765 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.316774 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.316789 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.316801 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.419944 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.419990 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.420001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.420022 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.420035 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.522646 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.522966 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.523004 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.523039 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.523049 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.625449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.625490 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.625499 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.625513 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.625522 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.728239 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.728275 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.728283 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.728299 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.728308 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.830573 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.830602 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.830610 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.830640 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.830650 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.932523 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.932579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.932596 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.932621 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:23 crc kubenswrapper[4987]: I1004 07:10:23.932661 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:23Z","lastTransitionTime":"2025-10-04T07:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.034829 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.034890 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.034908 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.034933 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.034951 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.053343 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:24 crc kubenswrapper[4987]: E1004 07:10:24.053517 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.072583 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.092392 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.109236 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.125698 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:24 crc kubenswrapper[4987]: E1004 07:10:24.125837 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:24 crc kubenswrapper[4987]: E1004 07:10:24.125891 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:10:28.125874613 +0000 UTC m=+45.180773455 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.130192 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.137580 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.137677 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.137699 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.137727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.137747 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.143925 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.160889 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.172187 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.186045 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.204933 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.223473 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.240195 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.240258 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.240272 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.240297 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.240312 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.243289 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.259865 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.275409 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.288029 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.309354 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.326595 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:24Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.344254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.344296 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.344306 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.344323 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.344334 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.446883 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.446932 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.446945 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.446967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.446982 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.549803 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.549864 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.549878 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.549899 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.549912 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.652826 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.652885 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.652901 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.652919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.652932 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.756265 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.756356 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.756373 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.756398 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.756417 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.859955 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.860014 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.860048 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.860078 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.860094 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.964150 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.964220 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.964239 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.964270 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:24 crc kubenswrapper[4987]: I1004 07:10:24.964291 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:24Z","lastTransitionTime":"2025-10-04T07:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.021143 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.021233 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.021254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.021290 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.021324 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.044842 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:25Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.051066 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.051152 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.051181 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.051213 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.051236 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.053302 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.053313 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.053441 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.053516 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.053416 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.053602 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.073669 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:25Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.079366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.079427 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.079445 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.079474 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.079492 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.100361 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:25Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.104931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.104986 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.105007 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.105027 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.105040 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.120207 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:25Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.125564 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.125614 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.125659 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.125688 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.125703 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.139596 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:25Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:25 crc kubenswrapper[4987]: E1004 07:10:25.139732 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.141954 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.142017 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.142030 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.142052 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.142062 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.244336 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.244390 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.244400 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.244416 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.244425 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.346684 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.346719 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.346728 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.346740 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.346749 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.449195 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.449228 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.449239 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.449267 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.449279 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.551967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.552016 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.552026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.552041 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.552058 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.658905 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.658991 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.659011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.659039 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.659067 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.761609 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.761696 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.761707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.761725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.761738 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.864849 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.864910 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.864925 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.864948 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.864964 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.967954 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.968012 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.968025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.968048 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:25 crc kubenswrapper[4987]: I1004 07:10:25.968073 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:25Z","lastTransitionTime":"2025-10-04T07:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.052956 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:26 crc kubenswrapper[4987]: E1004 07:10:26.053103 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.069779 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.069815 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.069825 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.069841 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.069853 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.172442 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.172503 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.172519 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.172540 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.172555 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.275431 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.275483 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.275495 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.275513 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.275526 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.379120 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.379167 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.379176 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.379193 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.379204 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.482144 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.482204 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.482218 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.482238 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.482252 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.586473 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.586535 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.586549 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.586582 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.586604 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.689810 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.689881 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.689898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.689935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.689954 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.793295 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.793382 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.793404 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.793434 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.793456 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.897201 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.897254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.897265 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.897284 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:26 crc kubenswrapper[4987]: I1004 07:10:26.897299 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:26Z","lastTransitionTime":"2025-10-04T07:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:26.999951 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.000015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.000030 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.000065 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.000083 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.053298 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.053387 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.053462 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:27 crc kubenswrapper[4987]: E1004 07:10:27.053460 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:27 crc kubenswrapper[4987]: E1004 07:10:27.053574 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:27 crc kubenswrapper[4987]: E1004 07:10:27.053737 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.103136 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.103200 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.103216 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.103239 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.103255 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.206022 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.206074 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.206087 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.206107 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.206126 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.309009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.309330 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.309441 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.309509 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.309587 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.411829 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.411887 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.411897 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.411915 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.411924 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.514503 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.514799 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.514893 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.514977 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.515060 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.617415 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.617451 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.617465 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.617490 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.617506 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.720641 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.720680 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.720698 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.720717 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.720728 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.823352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.823395 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.823414 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.823430 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.823442 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.926297 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.926341 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.926358 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.926380 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:27 crc kubenswrapper[4987]: I1004 07:10:27.926396 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:27Z","lastTransitionTime":"2025-10-04T07:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.029867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.029907 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.029925 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.029942 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.029953 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.053385 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:28 crc kubenswrapper[4987]: E1004 07:10:28.053584 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.131984 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.132025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.132040 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.132056 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.132065 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.168415 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:28 crc kubenswrapper[4987]: E1004 07:10:28.168546 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:28 crc kubenswrapper[4987]: E1004 07:10:28.168601 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:10:36.168588041 +0000 UTC m=+53.223486883 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.234215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.234254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.234263 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.234278 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.234290 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.336790 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.336828 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.336839 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.336854 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.336866 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.439637 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.439697 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.439707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.439721 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.439730 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.542069 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.542110 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.542126 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.542142 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.542150 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.644413 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.644443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.644453 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.644467 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.644498 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.746867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.746923 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.746939 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.746959 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.746973 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.848608 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.848661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.848673 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.848685 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.848694 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.951919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.951969 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.951978 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.951995 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:28 crc kubenswrapper[4987]: I1004 07:10:28.952010 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:28Z","lastTransitionTime":"2025-10-04T07:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.052578 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:29 crc kubenswrapper[4987]: E1004 07:10:29.052817 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.052865 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.052999 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:29 crc kubenswrapper[4987]: E1004 07:10:29.053484 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.053500 4987 scope.go:117] "RemoveContainer" containerID="eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2" Oct 04 07:10:29 crc kubenswrapper[4987]: E1004 07:10:29.053988 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.054176 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.054192 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.054202 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.054215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.054225 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.156768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.156817 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.156830 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.156846 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.156856 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.259147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.259195 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.259207 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.259222 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.259231 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.329711 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/1.log" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.331726 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.332204 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.348483 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.362186 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.362507 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.362522 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.362546 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.362555 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.362846 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.374173 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.387944 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.397476 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.407373 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.418770 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.434512 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.449503 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.465813 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.465904 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.465923 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.465952 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.465974 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.468218 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.481192 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.498995 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.514868 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.525981 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.546005 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.556988 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:29Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.568328 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.568358 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.568368 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.568383 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.568394 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.671257 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.671303 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.671316 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.671337 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.671350 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.773754 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.773798 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.773811 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.773827 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.773840 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.876053 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.876088 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.876102 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.876117 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.876128 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.978105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.978137 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.978146 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.978161 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:29 crc kubenswrapper[4987]: I1004 07:10:29.978169 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:29Z","lastTransitionTime":"2025-10-04T07:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.052948 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:30 crc kubenswrapper[4987]: E1004 07:10:30.053081 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.080164 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.080207 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.080227 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.080261 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.080274 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.183009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.183038 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.183047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.183092 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.183103 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.285776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.285837 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.285853 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.285872 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.285883 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.337665 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/2.log" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.338728 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/1.log" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.341578 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc" exitCode=1 Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.341657 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.341735 4987 scope.go:117] "RemoveContainer" containerID="eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.342438 4987 scope.go:117] "RemoveContainer" containerID="753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc" Oct 04 07:10:30 crc kubenswrapper[4987]: E1004 07:10:30.342683 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.360991 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.383117 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb2ef32e2ea898f7cdbd6b404e839c400ad24bd79bbc44adf9b774937cbc44c2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:17Z\\\",\\\"message\\\":\\\"enshift-controller-manager-operator,},ClusterIP:10.217.5.58,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.58],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nF1004 07:10:16.985319 6399 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for namespace Informer during admin network policy controller initialization, handler {0x1fcbf20 0x1fcbc00 0x1fcbba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: cur\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.387859 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.387898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.387913 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.387931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.387941 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.397667 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.415233 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.425159 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.438376 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.450106 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.462860 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.481330 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.490421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.490453 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.490469 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.490499 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.490512 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.493288 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.507035 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.519815 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.533115 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.542814 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.553437 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.564774 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.592542 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.592591 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.592603 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.592643 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.592655 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.694871 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.694937 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.694950 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.694969 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.694983 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.797661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.797977 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.798059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.798144 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.798216 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.900456 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.900492 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.900500 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.900518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:30 crc kubenswrapper[4987]: I1004 07:10:30.900527 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:30Z","lastTransitionTime":"2025-10-04T07:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.002306 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.002342 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.002352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.002366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.002376 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.053470 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.053525 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:31 crc kubenswrapper[4987]: E1004 07:10:31.053768 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.053530 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:31 crc kubenswrapper[4987]: E1004 07:10:31.054000 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:31 crc kubenswrapper[4987]: E1004 07:10:31.054439 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.104967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.105000 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.105009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.105022 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.105030 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.208016 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.208063 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.208074 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.208091 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.208107 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.310011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.310049 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.310058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.310074 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.310084 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.347063 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/2.log" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.350233 4987 scope.go:117] "RemoveContainer" containerID="753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc" Oct 04 07:10:31 crc kubenswrapper[4987]: E1004 07:10:31.350399 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.362792 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.377431 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.387902 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.396165 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.408125 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.411580 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.411634 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.411653 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.411674 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.411684 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.420120 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.430690 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.443435 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.458958 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.469476 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.479049 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.488041 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.503305 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.513954 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.514004 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.514017 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.514037 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.514050 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.516171 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.533399 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.551804 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:31Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.615844 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.615883 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.615893 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.615907 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.615916 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.718618 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.718695 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.718707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.718727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.718741 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.820659 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.820699 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.820709 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.820724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.820735 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.923437 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.923501 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.923519 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.923545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:31 crc kubenswrapper[4987]: I1004 07:10:31.923564 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:31Z","lastTransitionTime":"2025-10-04T07:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.026093 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.026180 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.026211 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.026238 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.026267 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.053539 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:32 crc kubenswrapper[4987]: E1004 07:10:32.053814 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.129496 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.129560 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.129579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.129601 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.129615 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.231858 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.231970 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.231989 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.232009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.232024 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.334849 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.334884 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.334896 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.334913 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.334926 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.438015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.438333 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.438352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.438373 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.438385 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.541844 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.542285 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.542560 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.542767 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.542833 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.635278 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.645602 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.645672 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.645686 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.645709 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.645724 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.655130 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.671239 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.692699 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.710410 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.723423 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.748117 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.748166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.748178 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.748201 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.748210 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.757293 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.773320 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.797220 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.811243 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.830419 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.844661 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.850360 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.850396 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.850407 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.850423 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.850433 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.857551 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.868414 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.880082 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.895695 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.910022 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:32Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.954047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.954102 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.954111 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.954125 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:32 crc kubenswrapper[4987]: I1004 07:10:32.954134 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:32Z","lastTransitionTime":"2025-10-04T07:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.052485 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.052485 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:33 crc kubenswrapper[4987]: E1004 07:10:33.052604 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:33 crc kubenswrapper[4987]: E1004 07:10:33.052694 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.052500 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:33 crc kubenswrapper[4987]: E1004 07:10:33.052772 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.056015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.056044 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.056053 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.056066 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.056076 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.159001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.159047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.159056 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.159071 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.159080 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.262496 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.262571 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.262591 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.262671 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.262715 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.365215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.365260 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.365269 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.365288 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.365300 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.468091 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.468156 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.468166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.468180 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.468190 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.570886 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.570935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.570945 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.570961 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.570974 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.672755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.672784 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.672794 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.672807 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.672815 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.775185 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.775233 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.775244 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.775263 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.775275 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.878275 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.878328 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.878337 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.878352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.878362 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.981266 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.981319 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.981340 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.981364 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:33 crc kubenswrapper[4987]: I1004 07:10:33.981381 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:33Z","lastTransitionTime":"2025-10-04T07:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.053078 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:34 crc kubenswrapper[4987]: E1004 07:10:34.053220 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.067973 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.084233 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.084290 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.084310 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.084336 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.084354 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.086161 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.105073 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.119546 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.138847 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.155384 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.168758 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.178986 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.187444 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.187474 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.187485 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.187502 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.187513 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.192799 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.209100 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.224360 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.240318 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.259537 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.290334 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.290431 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.290454 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.290525 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.290435 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.290816 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.309167 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.340835 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:34Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.394024 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.394094 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.394118 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.394148 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.394170 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.498938 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.498987 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.499000 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.499019 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.499033 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.601403 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.601449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.601461 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.601482 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.601496 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.704341 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.704400 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.704421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.704444 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.704461 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.806783 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.806849 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.806861 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.806876 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.806886 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.909472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.909517 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.909526 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.909542 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:34 crc kubenswrapper[4987]: I1004 07:10:34.909552 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:34Z","lastTransitionTime":"2025-10-04T07:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.012122 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.012164 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.012173 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.012187 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.012196 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.053008 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.053094 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.053154 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.053021 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.053305 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.053427 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.115093 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.115180 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.115203 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.115236 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.115259 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.217007 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.217042 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.217051 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.217064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.217074 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.319377 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.319440 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.319449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.319462 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.319471 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.421427 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.421465 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.421473 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.421486 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.421495 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.499773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.499873 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.499884 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.499898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.499907 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.514826 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:35Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.518381 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.518415 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.518424 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.518440 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.518451 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.529831 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:35Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.532573 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.532809 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.532891 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.532918 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.532931 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.546561 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:35Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.549914 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.549953 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.549964 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.549982 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.549994 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.562366 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:35Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.565371 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.565423 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.565435 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.565455 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.565467 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.579920 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:35Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:35 crc kubenswrapper[4987]: E1004 07:10:35.580042 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.581518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.581547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.581557 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.581570 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.581579 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.683114 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.683161 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.683176 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.683193 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.683203 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.785333 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.785388 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.785400 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.785415 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.785425 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.887981 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.888040 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.888070 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.888098 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.888118 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.989974 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.989999 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.990006 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.990018 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:35 crc kubenswrapper[4987]: I1004 07:10:35.990026 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:35Z","lastTransitionTime":"2025-10-04T07:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.053294 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.053495 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.091533 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.091572 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.091582 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.091594 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.091603 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.193959 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.193989 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.193998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.194011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.194021 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.252894 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.253048 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.253103 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:10:52.253090438 +0000 UTC m=+69.307989280 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.296302 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.296337 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.296347 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.296362 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.296372 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.397851 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.397895 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.397906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.397920 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.397932 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.500254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.500306 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.500314 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.500328 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.500337 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.602460 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.602509 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.602525 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.602544 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.602558 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.704872 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.704919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.704931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.704947 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.704959 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.756546 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.756811 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:11:08.756776493 +0000 UTC m=+85.811675345 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.807389 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.807421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.807430 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.807443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.807452 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.857300 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.857387 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.857427 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857462 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857517 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857538 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:11:08.857518642 +0000 UTC m=+85.912417524 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857539 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857560 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.857470 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857595 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857597 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:11:08.857585935 +0000 UTC m=+85.912484827 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857692 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857861 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857888 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857905 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:11:08.857786181 +0000 UTC m=+85.912685023 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:10:36 crc kubenswrapper[4987]: E1004 07:10:36.857973 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:11:08.857944175 +0000 UTC m=+85.912843137 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.910463 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.910510 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.910524 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.910546 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:36 crc kubenswrapper[4987]: I1004 07:10:36.910561 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:36Z","lastTransitionTime":"2025-10-04T07:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.013526 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.013608 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.013643 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.013662 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.013674 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.053154 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.053198 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.053314 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:37 crc kubenswrapper[4987]: E1004 07:10:37.053468 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:37 crc kubenswrapper[4987]: E1004 07:10:37.053672 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:37 crc kubenswrapper[4987]: E1004 07:10:37.053914 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.116539 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.116583 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.116593 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.116614 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.116647 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.219765 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.219827 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.219838 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.219858 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.219873 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.322496 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.322555 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.322574 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.322602 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.322650 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.324211 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.336607 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.342071 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.362712 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.378446 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.398010 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.415789 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.426198 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.426250 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.426264 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.426285 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.426300 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.430774 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.448938 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.465723 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.482029 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.495994 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.511234 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.526771 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.528155 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.528199 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.528215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.528235 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.528246 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.541866 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.553466 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.568577 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.585124 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:37Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.663709 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.663754 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.663765 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.663783 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.663795 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.767037 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.767077 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.767090 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.767107 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.767118 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.869311 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.869350 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.869361 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.869374 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.869382 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.973184 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.973214 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.973222 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.973234 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:37 crc kubenswrapper[4987]: I1004 07:10:37.973243 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:37Z","lastTransitionTime":"2025-10-04T07:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.052956 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:38 crc kubenswrapper[4987]: E1004 07:10:38.053309 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.075225 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.075472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.075551 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.075668 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.075754 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.178217 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.178253 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.178264 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.178281 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.178292 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.281115 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.281156 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.281188 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.281205 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.281216 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.382786 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.382824 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.382834 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.382848 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.382858 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.484895 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.484934 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.484944 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.484962 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.484974 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.586954 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.587009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.587018 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.587033 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.587043 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.688844 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.688903 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.688914 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.688931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.688942 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.791724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.791803 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.791832 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.791863 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.791886 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.894020 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.894077 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.894090 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.894107 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.894119 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.996419 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.996466 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.996477 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.996494 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:38 crc kubenswrapper[4987]: I1004 07:10:38.996512 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:38Z","lastTransitionTime":"2025-10-04T07:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.053158 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.053181 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:39 crc kubenswrapper[4987]: E1004 07:10:39.053297 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:39 crc kubenswrapper[4987]: E1004 07:10:39.053363 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.053197 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:39 crc kubenswrapper[4987]: E1004 07:10:39.053435 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.099100 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.099138 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.099147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.099163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.099172 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.201437 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.201495 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.201507 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.201525 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.201534 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.303485 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.303535 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.303550 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.303571 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.303586 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.408661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.408796 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.408809 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.408825 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.408836 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.510863 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.510897 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.510928 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.510945 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.510957 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.613498 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.613588 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.613616 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.613649 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.613660 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.715736 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.715989 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.716050 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.716151 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.716214 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.818719 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.818791 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.818808 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.818833 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.818850 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.920823 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.920866 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.920876 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.920907 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:39 crc kubenswrapper[4987]: I1004 07:10:39.920921 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:39Z","lastTransitionTime":"2025-10-04T07:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.023807 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.023846 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.023855 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.023870 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.023881 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.053317 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:40 crc kubenswrapper[4987]: E1004 07:10:40.053450 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.125916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.125956 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.125967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.125983 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.125994 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.227964 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.227999 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.228009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.228026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.228036 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.330365 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.330424 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.330434 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.330449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.330459 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.432496 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.432530 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.432538 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.432550 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.432561 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.534704 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.534755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.534764 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.534778 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.534788 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.637246 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.637274 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.637282 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.637293 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.637302 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.739769 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.739831 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.739844 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.739860 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.739871 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.842743 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.842792 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.842803 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.842834 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.842845 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.945947 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.945996 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.946007 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.946023 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:40 crc kubenswrapper[4987]: I1004 07:10:40.946035 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:40Z","lastTransitionTime":"2025-10-04T07:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.048579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.048655 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.048664 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.048678 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.048691 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.052904 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.052961 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.052968 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:41 crc kubenswrapper[4987]: E1004 07:10:41.053004 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:41 crc kubenswrapper[4987]: E1004 07:10:41.053063 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:41 crc kubenswrapper[4987]: E1004 07:10:41.053133 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.151410 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.151458 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.151466 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.151480 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.151493 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.253724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.253766 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.253781 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.253797 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.253818 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.356078 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.356116 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.356126 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.356142 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.356152 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.458457 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.458496 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.458505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.458519 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.458529 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.560791 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.560829 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.560840 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.560854 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.560864 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.663316 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.663367 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.663402 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.663424 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.663437 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.766031 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.766067 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.766074 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.766088 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.766101 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.868731 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.868824 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.868851 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.868881 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.868902 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.972001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.972040 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.972051 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.972069 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:41 crc kubenswrapper[4987]: I1004 07:10:41.972080 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:41Z","lastTransitionTime":"2025-10-04T07:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.053483 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:42 crc kubenswrapper[4987]: E1004 07:10:42.054367 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.074805 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.074854 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.074864 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.074881 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.074890 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.177933 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.177995 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.178007 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.178028 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.178042 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.280755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.280919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.280961 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.280987 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.281004 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.383011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.383058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.383068 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.383086 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.383098 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.486307 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.486372 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.486395 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.486426 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.486449 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.588884 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.588976 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.589004 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.589035 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.589059 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.691843 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.691880 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.691891 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.691906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.691917 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.794103 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.794193 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.794208 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.794226 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.794236 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.896939 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.896985 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.896998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.897015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.897027 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.999657 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.999684 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.999691 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.999706 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:42 crc kubenswrapper[4987]: I1004 07:10:42.999714 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:42Z","lastTransitionTime":"2025-10-04T07:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.052435 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.052454 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.052435 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:43 crc kubenswrapper[4987]: E1004 07:10:43.052562 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:43 crc kubenswrapper[4987]: E1004 07:10:43.052615 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:43 crc kubenswrapper[4987]: E1004 07:10:43.052748 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.102252 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.102315 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.102329 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.102353 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.102368 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.204114 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.204207 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.204238 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.204271 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.204291 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.306666 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.306712 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.306724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.306741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.306751 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.409166 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.409202 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.409210 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.409224 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.409233 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.511545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.511586 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.511598 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.511614 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.511649 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.614683 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.614740 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.614757 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.614777 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.614792 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.717274 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.717316 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.717326 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.717342 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.717353 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.820808 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.820853 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.820862 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.820879 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.820889 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.923060 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.923094 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.923103 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.923115 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:43 crc kubenswrapper[4987]: I1004 07:10:43.923125 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:43Z","lastTransitionTime":"2025-10-04T07:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.025885 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.025934 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.025945 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.025983 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.025997 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.052808 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:44 crc kubenswrapper[4987]: E1004 07:10:44.053016 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.068697 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.081396 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.095016 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.106447 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.115726 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.127651 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.128543 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.128578 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.128607 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.128660 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.128673 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.142923 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.153927 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.163980 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.177568 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.189446 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.205669 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.215965 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.228894 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.230724 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.230750 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.230759 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.230773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.230784 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.238435 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.248947 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.260428 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:44Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.333018 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.333053 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.333061 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.333075 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.333083 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.435939 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.435974 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.435983 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.435996 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.436005 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.539012 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.539391 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.539421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.539449 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.539484 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.641736 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.641779 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.641790 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.641807 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.641819 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.744642 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.744688 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.744701 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.744719 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.744731 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.847053 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.847084 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.847296 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.847309 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.847317 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.950351 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.950393 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.950402 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.950416 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:44 crc kubenswrapper[4987]: I1004 07:10:44.950425 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:44Z","lastTransitionTime":"2025-10-04T07:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052551 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052551 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052641 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052651 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052666 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052675 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.052769 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052764 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.052824 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.053022 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.053098 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.053319 4987 scope.go:117] "RemoveContainer" containerID="753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc" Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.053520 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.155843 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.155880 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.155890 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.155903 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.155914 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.259309 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.259357 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.259370 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.259386 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.259395 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.362455 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.362507 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.362523 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.362546 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.362562 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.464984 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.465038 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.465049 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.465124 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.465148 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.567728 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.567777 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.567789 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.567806 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.567817 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.670243 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.670293 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.670301 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.670317 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.670326 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.686118 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.686154 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.686163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.686182 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.686191 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.701538 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:45Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.705741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.705777 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.705785 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.705799 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.705807 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.723468 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:45Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.727364 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.727402 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.727414 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.727431 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.727445 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.744279 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:45Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.747457 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.747493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.747505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.747521 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.747532 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.763354 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:45Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.766371 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.766405 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.766417 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.766429 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.766437 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.777694 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:45Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:45 crc kubenswrapper[4987]: E1004 07:10:45.777962 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.780034 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.780100 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.780121 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.780148 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.780165 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.882965 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.883006 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.883015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.883029 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.883038 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.986309 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.986346 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.986355 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.986368 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:45 crc kubenswrapper[4987]: I1004 07:10:45.986379 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:45Z","lastTransitionTime":"2025-10-04T07:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.053062 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:46 crc kubenswrapper[4987]: E1004 07:10:46.053218 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.089912 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.089983 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.090005 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.090036 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.090057 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.193493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.193545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.193561 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.193586 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.193602 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.296411 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.296443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.296450 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.296463 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.296472 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.399111 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.399167 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.399187 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.399214 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.399239 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.502255 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.502295 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.502306 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.502325 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.502337 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.605980 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.606030 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.606047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.606072 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.606090 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.709780 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.709831 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.709844 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.709865 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.709880 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.812703 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.812761 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.812775 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.812802 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.812817 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.915221 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.915257 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.915267 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.915285 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:46 crc kubenswrapper[4987]: I1004 07:10:46.915295 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:46Z","lastTransitionTime":"2025-10-04T07:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.017856 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.017902 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.017916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.017939 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.017956 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.052758 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:47 crc kubenswrapper[4987]: E1004 07:10:47.052873 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.053036 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:47 crc kubenswrapper[4987]: E1004 07:10:47.053076 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.053174 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:47 crc kubenswrapper[4987]: E1004 07:10:47.053214 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.121421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.121457 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.121467 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.121482 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.121495 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.223655 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.223684 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.223696 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.223711 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.223722 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.327946 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.327977 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.327987 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.328003 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.328014 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.431128 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.431165 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.431174 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.431189 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.431198 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.532951 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.532979 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.532988 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.533001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.533010 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.635560 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.635647 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.635661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.635682 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.635693 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.756661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.756745 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.756760 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.756787 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.756802 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.858680 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.858714 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.858727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.858742 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.858755 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.960978 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.961027 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.961041 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.961058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:47 crc kubenswrapper[4987]: I1004 07:10:47.961073 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:47Z","lastTransitionTime":"2025-10-04T07:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.053467 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:48 crc kubenswrapper[4987]: E1004 07:10:48.053695 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.071093 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.071135 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.071147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.071169 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.071190 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.173579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.173673 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.173688 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.173733 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.173760 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.276016 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.276087 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.276105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.276147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.276161 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.379016 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.379065 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.379082 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.379518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.379550 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.486607 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.486739 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.486752 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.486768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.486779 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.588898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.588935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.588946 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.588963 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.588973 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.690916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.690960 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.690969 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.690984 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.690994 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.793508 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.793554 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.793565 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.793579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.793589 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.895963 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.895993 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.896001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.896015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.896023 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.998211 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.998503 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.998572 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.998680 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:48 crc kubenswrapper[4987]: I1004 07:10:48.998753 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:48Z","lastTransitionTime":"2025-10-04T07:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.052803 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.052832 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.052836 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:49 crc kubenswrapper[4987]: E1004 07:10:49.052926 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:49 crc kubenswrapper[4987]: E1004 07:10:49.053034 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:49 crc kubenswrapper[4987]: E1004 07:10:49.053172 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.101437 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.101475 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.101487 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.101504 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.101515 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.204532 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.204592 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.204604 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.204648 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.204667 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.307263 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.307319 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.307335 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.307356 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.307368 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.410114 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.410152 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.410161 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.410174 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.410217 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.513352 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.513402 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.513411 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.513426 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.513436 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.616079 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.616116 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.616125 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.616138 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.616147 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.719114 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.719208 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.719224 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.719250 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.719267 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.821481 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.821548 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.821562 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.821581 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.821592 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.923678 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.923723 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.923735 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.923751 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:49 crc kubenswrapper[4987]: I1004 07:10:49.923762 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:49Z","lastTransitionTime":"2025-10-04T07:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.026662 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.026700 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.026711 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.026726 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.026737 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.053188 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:50 crc kubenswrapper[4987]: E1004 07:10:50.053323 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.078311 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.128961 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.129002 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.129012 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.129025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.129035 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.231144 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.231185 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.231195 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.231213 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.231223 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.333609 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.333698 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.333710 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.333725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.333734 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.435489 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.435529 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.435538 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.435554 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.435564 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.538212 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.538254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.538264 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.538278 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.538289 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.640957 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.640999 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.641026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.641045 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.641060 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.742892 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.742931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.742939 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.742959 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.742968 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.845630 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.845675 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.845690 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.845707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.845719 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.948697 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.948746 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.948755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.948768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:50 crc kubenswrapper[4987]: I1004 07:10:50.948776 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:50Z","lastTransitionTime":"2025-10-04T07:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.051350 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.051382 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.051391 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.051405 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.051413 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.052579 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.052618 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.052640 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:51 crc kubenswrapper[4987]: E1004 07:10:51.052740 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:51 crc kubenswrapper[4987]: E1004 07:10:51.052793 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:51 crc kubenswrapper[4987]: E1004 07:10:51.052860 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.156794 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.156846 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.156858 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.156875 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.156897 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.260946 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.260995 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.261005 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.261020 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.261031 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.363789 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.363838 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.363855 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.363879 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.363897 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.466727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.466756 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.466767 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.466784 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.466795 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.569493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.569526 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.569538 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.569554 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.569567 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.671967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.672008 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.672018 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.672037 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.672048 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.774246 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.774280 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.774288 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.774301 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.774311 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.876249 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.876281 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.876291 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.876304 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.876313 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.979038 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.979101 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.979116 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.979140 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:51 crc kubenswrapper[4987]: I1004 07:10:51.979157 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:51Z","lastTransitionTime":"2025-10-04T07:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.053517 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:52 crc kubenswrapper[4987]: E1004 07:10:52.053798 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.081703 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.081753 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.081766 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.081785 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.081822 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.184704 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.184757 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.184772 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.184797 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.184820 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.286738 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.286770 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.286778 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.286790 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.286799 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.310929 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:52 crc kubenswrapper[4987]: E1004 07:10:52.311073 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:52 crc kubenswrapper[4987]: E1004 07:10:52.311122 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:11:24.311106612 +0000 UTC m=+101.366005464 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.388757 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.388805 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.388816 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.388834 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.388846 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.490739 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.490792 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.490804 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.490819 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.490831 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.593190 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.593232 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.593246 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.593263 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.593275 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.695768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.696269 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.696341 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.696443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.696502 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.799097 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.799136 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.799146 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.799160 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.799171 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.901300 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.901345 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.901357 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.901372 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:52 crc kubenswrapper[4987]: I1004 07:10:52.901383 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:52Z","lastTransitionTime":"2025-10-04T07:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.003868 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.003935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.003952 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.004015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.004033 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.052495 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.052509 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.052520 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:53 crc kubenswrapper[4987]: E1004 07:10:53.052738 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:53 crc kubenswrapper[4987]: E1004 07:10:53.052824 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:53 crc kubenswrapper[4987]: E1004 07:10:53.052915 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.106254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.106290 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.106298 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.106311 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.106320 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.208665 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.208725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.208739 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.208755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.208789 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.313518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.313560 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.313570 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.313585 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.313598 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.414774 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.414814 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.414822 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.414837 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.414848 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.517778 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.517824 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.517837 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.517855 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.517867 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.621057 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.621097 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.621105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.621119 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.621130 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.723437 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.723493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.723507 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.723522 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.723533 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.826199 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.826257 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.826270 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.826283 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.826292 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.932196 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.932254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.932267 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.932284 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:53 crc kubenswrapper[4987]: I1004 07:10:53.932296 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:53Z","lastTransitionTime":"2025-10-04T07:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.034935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.034986 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.035001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.035022 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.035038 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.053519 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:54 crc kubenswrapper[4987]: E1004 07:10:54.053659 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.089063 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.112956 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.128110 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.136198 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.136225 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.136234 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.136245 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.136255 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.146215 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.157435 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.171280 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.182474 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.193768 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.205252 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.220562 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.232461 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.238067 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.238099 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.238111 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.238139 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.238148 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.245188 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.256102 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.268926 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.279556 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.291443 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.303658 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.315674 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.340518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.340553 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.340565 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.340580 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.340591 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.418135 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/0.log" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.418181 4987 generic.go:334] "Generic (PLEG): container finished" podID="659797bb-8ec1-4ad8-880e-7aec3b068c46" containerID="d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599" exitCode=1 Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.418209 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerDied","Data":"d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.418539 4987 scope.go:117] "RemoveContainer" containerID="d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.432248 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.442873 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.442910 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.442920 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.442938 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.442949 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.444864 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.457888 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.470143 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.484306 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.496211 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.505070 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.513961 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.524499 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.535118 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.544977 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.545011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.545022 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.545037 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.545049 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.546679 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.565198 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.577289 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.588780 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.601092 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.611284 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.625141 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.638236 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:54Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.646768 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.646792 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.646800 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.646814 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.646823 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.748717 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.748752 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.748760 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.748776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.748791 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.850886 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.850916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.850924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.850936 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.850945 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.953089 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.953118 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.953129 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.953144 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:54 crc kubenswrapper[4987]: I1004 07:10:54.953155 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:54Z","lastTransitionTime":"2025-10-04T07:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.053367 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:55 crc kubenswrapper[4987]: E1004 07:10:55.053483 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.053689 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:55 crc kubenswrapper[4987]: E1004 07:10:55.053793 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.053696 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:55 crc kubenswrapper[4987]: E1004 07:10:55.053913 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.055027 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.055053 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.055090 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.055103 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.055115 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.156941 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.156984 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.156994 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.157009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.157022 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.259288 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.259331 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.259343 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.259359 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.259371 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.361436 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.361472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.361480 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.361493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.361502 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.423372 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/0.log" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.423433 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerStarted","Data":"d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.435786 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.448930 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.459662 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.462889 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.462916 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.462924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.462936 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.462944 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.477594 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.488477 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.498202 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.507754 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.520466 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.529972 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.539571 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.548386 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.559055 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.564978 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.565006 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.565015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.565028 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.565037 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.572737 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.584314 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.596760 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.606977 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.617215 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.626529 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:55Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.667404 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.667444 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.667454 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.667470 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.667480 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.769295 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.769335 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.769346 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.769396 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.769409 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.871520 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.871559 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.871569 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.871582 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.871592 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.973509 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.973541 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.973554 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.973568 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.973578 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.997707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.997730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.997741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.997755 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:55 crc kubenswrapper[4987]: I1004 07:10:55.997765 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:55Z","lastTransitionTime":"2025-10-04T07:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.007885 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.010588 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.010637 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.010653 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.010672 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.010684 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.021004 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.023860 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.023885 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.023893 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.023907 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.023915 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.036116 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.039336 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.039376 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.039388 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.039403 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.039416 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.052149 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.052560 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.052682 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.055769 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.055807 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.055819 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.055835 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.055846 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.067672 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:56Z is after 2025-08-24T17:21:41Z" Oct 04 07:10:56 crc kubenswrapper[4987]: E1004 07:10:56.067888 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.076233 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.076291 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.076313 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.076342 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.076359 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.178874 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.178914 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.178924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.178940 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.178951 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.280954 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.280982 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.280991 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.281004 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.281012 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.383569 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.383613 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.383638 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.383656 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.383667 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.485781 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.485819 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.485828 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.485843 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.485853 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.587959 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.588000 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.588012 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.588029 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.588041 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.690539 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.690583 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.690598 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.690618 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.690655 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.793011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.793047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.793058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.793074 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.793085 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.895472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.895513 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.895733 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.895762 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.895788 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.998059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.998094 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.998105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.998120 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:56 crc kubenswrapper[4987]: I1004 07:10:56.998131 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:56Z","lastTransitionTime":"2025-10-04T07:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.052609 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:57 crc kubenswrapper[4987]: E1004 07:10:57.052777 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.052812 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:57 crc kubenswrapper[4987]: E1004 07:10:57.052950 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.053154 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:57 crc kubenswrapper[4987]: E1004 07:10:57.053232 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.100609 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.100688 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.100702 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.100725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.100740 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.203839 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.203892 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.203903 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.203920 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.203931 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.306578 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.306649 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.306661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.306677 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.306688 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.409210 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.409245 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.409254 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.409268 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.409280 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.512575 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.512644 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.512657 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.512678 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.512688 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.615637 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.615679 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.615689 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.615704 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.615717 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.718262 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.718312 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.718323 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.718340 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.718351 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.821025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.821112 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.821139 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.821162 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.821179 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.924655 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.924731 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.924746 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.924769 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:57 crc kubenswrapper[4987]: I1004 07:10:57.924783 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:57Z","lastTransitionTime":"2025-10-04T07:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.028071 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.028134 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.028147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.028172 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.028183 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.053250 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:10:58 crc kubenswrapper[4987]: E1004 07:10:58.053507 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.131546 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.131605 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.131617 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.131683 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.131696 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.236013 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.236049 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.236058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.236072 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.236080 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.339341 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.339415 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.339433 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.339459 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.339475 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.442391 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.442463 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.442472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.442500 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.442512 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.544810 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.544848 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.544857 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.544871 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.544881 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.646830 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.646875 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.646884 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.646900 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.646911 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.749551 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.749591 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.749601 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.749615 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.749659 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.852476 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.852550 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.852564 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.852590 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.852604 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.955163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.955247 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.955275 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.955305 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:58 crc kubenswrapper[4987]: I1004 07:10:58.955323 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:58Z","lastTransitionTime":"2025-10-04T07:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.052863 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.052975 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.052863 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:10:59 crc kubenswrapper[4987]: E1004 07:10:59.053072 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:10:59 crc kubenswrapper[4987]: E1004 07:10:59.053261 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:10:59 crc kubenswrapper[4987]: E1004 07:10:59.053308 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.057446 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.057481 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.057490 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.057504 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.057514 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.159613 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.159675 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.159685 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.159698 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.159706 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.262953 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.262998 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.263009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.263025 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.263037 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.365461 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.365510 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.365522 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.365540 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.365552 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.469435 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.469528 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.469553 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.469589 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.469619 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.572844 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.572906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.572919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.572938 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.572950 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.675756 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.675799 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.675810 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.675828 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.675840 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.778868 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.778918 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.778930 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.778950 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.778965 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.882479 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.882557 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.882571 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.882598 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.882618 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.986348 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.986401 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.986412 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.986432 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:10:59 crc kubenswrapper[4987]: I1004 07:10:59.986444 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:10:59Z","lastTransitionTime":"2025-10-04T07:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.053384 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:00 crc kubenswrapper[4987]: E1004 07:11:00.053546 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.054460 4987 scope.go:117] "RemoveContainer" containerID="753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.089472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.089500 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.089508 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.089521 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.089531 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.192439 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.192496 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.192513 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.192538 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.192557 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.294716 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.294776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.294788 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.294805 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.294818 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.397522 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.397559 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.397568 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.397586 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.397599 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.445249 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/2.log" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.448927 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.449669 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.461866 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.474302 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.486107 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.497132 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.499510 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.499545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.499556 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.499572 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.499584 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.510886 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.533162 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.547062 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.562978 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.575713 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.601205 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.602794 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.602840 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.602851 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.602871 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.602910 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.617799 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.637949 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.654911 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.667824 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.677195 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.690015 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.701998 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.705976 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.706015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.706026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.706044 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.706056 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.714243 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:00Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.808251 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.808285 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.808295 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.808308 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.808317 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.910729 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.910811 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.910824 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.910842 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:00 crc kubenswrapper[4987]: I1004 07:11:00.910856 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:00Z","lastTransitionTime":"2025-10-04T07:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.012833 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.012882 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.012891 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.012924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.012935 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.053379 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.053439 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:01 crc kubenswrapper[4987]: E1004 07:11:01.053495 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.053452 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:01 crc kubenswrapper[4987]: E1004 07:11:01.053644 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:01 crc kubenswrapper[4987]: E1004 07:11:01.053701 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.114664 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.114707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.114747 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.114762 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.114770 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.217563 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.217615 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.217643 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.217659 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.217672 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.319820 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.319854 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.319864 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.319877 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.319887 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.422822 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.422891 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.422903 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.422922 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.422936 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.454743 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/3.log" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.455560 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/2.log" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.458759 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" exitCode=1 Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.458814 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.458874 4987 scope.go:117] "RemoveContainer" containerID="753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.459990 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:11:01 crc kubenswrapper[4987]: E1004 07:11:01.460298 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.472313 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.485950 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.497640 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.514054 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.525533 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.525576 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.525584 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.525600 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.525610 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.531827 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.548377 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.560709 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.580115 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.590393 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.608125 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.624178 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.627876 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.627967 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.627986 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.628014 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.628031 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.644784 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.675466 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://753438c33dada5375e574149950ce1f5645e002556a4939728d63db4608c33cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:30Z\\\",\\\"message\\\":\\\": Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:10:30Z is after 2025-08-24T17:21:41Z]\\\\nI1004 07:10:30.081308 6636 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm after 0 failed attempt(s)\\\\nI1004 07:10:30.081778 6636 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:10:30.081387 6636 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081833 6636 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1004 07:10:30.081815 6636 model_\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:11:00Z\\\",\\\"message\\\":\\\"5ca3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:11:00.942941 7018 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"97419c58-41c7-41d7-a137-a446f0c7eeb3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:11:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.688789 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.701237 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.712266 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.728696 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.730338 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.730389 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.730403 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.730425 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.730437 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.740311 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:01Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.832658 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.832708 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.832720 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.832743 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.832755 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.935194 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.935266 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.935287 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.935313 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:01 crc kubenswrapper[4987]: I1004 07:11:01.935331 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:01Z","lastTransitionTime":"2025-10-04T07:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.039842 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.039922 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.039950 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.039981 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.040002 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.053295 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:02 crc kubenswrapper[4987]: E1004 07:11:02.053481 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.141835 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.141901 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.141919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.141942 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.141960 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.244543 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.244586 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.244598 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.244616 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.244661 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.348197 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.348262 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.348278 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.348296 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.348308 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.450696 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.450727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.450734 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.450747 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.450756 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.467896 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/3.log" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.472440 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:11:02 crc kubenswrapper[4987]: E1004 07:11:02.472805 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.489229 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.506860 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.524108 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.540057 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.552997 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.553059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.553079 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.553103 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.553122 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.556101 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.567886 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.580989 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.601436 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.618403 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.634543 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.655378 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.655835 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.655871 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.655883 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.655904 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.655917 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.670021 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.703376 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:11:00Z\\\",\\\"message\\\":\\\"5ca3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:11:00.942941 7018 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"97419c58-41c7-41d7-a137-a446f0c7eeb3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:11:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.722955 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.736577 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.758662 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.758708 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.758719 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.758736 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.758747 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.761129 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.774001 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.789288 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:02Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.861363 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.861412 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.861421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.861438 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.861447 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.963991 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.964064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.964082 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.964104 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:02 crc kubenswrapper[4987]: I1004 07:11:02.964121 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:02Z","lastTransitionTime":"2025-10-04T07:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.052708 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.052747 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.052716 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:03 crc kubenswrapper[4987]: E1004 07:11:03.052941 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:03 crc kubenswrapper[4987]: E1004 07:11:03.053023 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:03 crc kubenswrapper[4987]: E1004 07:11:03.053145 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.066860 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.066904 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.066915 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.066935 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.066952 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.171657 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.171721 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.171743 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.171773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.171794 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.275105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.275162 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.275174 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.275192 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.275204 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.378847 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.378897 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.378906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.378926 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.378936 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.481606 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.481697 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.481712 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.481736 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.481748 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.583906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.583958 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.583968 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.583981 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.583991 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.686228 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.686267 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.686275 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.686291 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.686303 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.788602 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.788654 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.788664 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.788680 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.788689 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.890479 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.890519 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.890529 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.890548 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.890564 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.992732 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.992766 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.992773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.992786 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:03 crc kubenswrapper[4987]: I1004 07:11:03.992795 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:03Z","lastTransitionTime":"2025-10-04T07:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.052683 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:04 crc kubenswrapper[4987]: E1004 07:11:04.052831 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.070125 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.082758 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.094958 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.094992 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.095001 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.095016 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.095026 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.096716 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.106539 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.114809 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.123996 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.135684 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.147652 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.159073 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.172089 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.187357 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.200153 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.200710 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.200741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.200750 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.200765 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.200774 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.219451 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:11:00Z\\\",\\\"message\\\":\\\"5ca3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:11:00.942941 7018 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"97419c58-41c7-41d7-a137-a446f0c7eeb3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:11:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.231697 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.245345 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.253906 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.264286 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.274583 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:04Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.303684 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.303729 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.303739 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.303754 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.303766 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.405579 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.405620 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.405648 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.405662 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.405675 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.508065 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.508097 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.508106 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.508119 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.508129 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.610708 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.610761 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.610775 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.610793 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.610804 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.713220 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.713274 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.713286 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.713303 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.713315 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.815163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.815195 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.815203 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.815215 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.815225 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.918850 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.918906 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.918923 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.918947 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:04 crc kubenswrapper[4987]: I1004 07:11:04.918964 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:04Z","lastTransitionTime":"2025-10-04T07:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.022174 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.022214 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.022226 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.022270 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.022283 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.052534 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.052570 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.052533 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:05 crc kubenswrapper[4987]: E1004 07:11:05.052781 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:05 crc kubenswrapper[4987]: E1004 07:11:05.053011 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:05 crc kubenswrapper[4987]: E1004 07:11:05.053136 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.125237 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.125294 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.125310 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.125334 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.125351 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.227198 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.227245 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.227259 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.227280 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.227294 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.330673 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.330763 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.330788 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.330820 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.330847 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.433129 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.433165 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.433172 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.433185 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.433195 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.535765 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.535819 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.535831 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.535848 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.535860 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.638216 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.638259 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.638270 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.638284 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.638295 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.741102 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.741145 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.741158 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.741173 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.741185 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.843882 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.844114 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.844126 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.844142 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.844152 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.946416 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.946458 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.946468 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.946482 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:05 crc kubenswrapper[4987]: I1004 07:11:05.946491 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:05Z","lastTransitionTime":"2025-10-04T07:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.048917 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.048953 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.048962 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.048978 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.048989 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.053814 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.053947 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.151159 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.151199 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.151217 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.151234 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.151244 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.205257 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.205289 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.205297 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.205309 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.205318 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.216966 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.221006 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.221056 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.221068 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.221112 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.221124 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.231637 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.234690 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.234753 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.234766 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.234783 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.234794 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.251093 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.255190 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.255241 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.255260 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.255283 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.255302 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.268600 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.271847 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.271895 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.271912 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.271936 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.271952 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.284206 4987 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T07:11:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4ae34cc5-3302-4911-8938-ea69ea878ab9\\\",\\\"systemUUID\\\":\\\"7dabc7f6-4527-406f-a679-876deca45080\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:06Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:06 crc kubenswrapper[4987]: E1004 07:11:06.284370 4987 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.286120 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.286147 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.286157 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.286172 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.286182 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.389122 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.389186 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.389198 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.389220 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.389232 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.491293 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.491333 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.491349 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.491366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.491376 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.593996 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.594046 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.594058 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.594076 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.594092 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.698367 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.698506 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.698518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.698532 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.698544 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.801227 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.801284 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.801306 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.801335 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.801356 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.905055 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.905433 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.905443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.905473 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:06 crc kubenswrapper[4987]: I1004 07:11:06.905483 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:06Z","lastTransitionTime":"2025-10-04T07:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.007448 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.007483 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.007499 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.007521 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.007533 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.052491 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.052610 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:07 crc kubenswrapper[4987]: E1004 07:11:07.053895 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:07 crc kubenswrapper[4987]: E1004 07:11:07.053926 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.054105 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:07 crc kubenswrapper[4987]: E1004 07:11:07.054527 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.111272 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.111518 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.111582 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.111665 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.111737 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.213993 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.214065 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.214083 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.214114 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.214133 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.317771 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.317817 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.317827 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.317845 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.317857 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.420523 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.420556 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.420565 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.420578 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.420590 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.523516 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.523564 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.523577 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.523595 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.523607 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.626034 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.626087 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.626111 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.626131 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.626146 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.728652 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.728744 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.728776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.728803 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.728825 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.831338 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.831380 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.831389 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.831406 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.831415 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.934081 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.934116 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.934124 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.934138 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:07 crc kubenswrapper[4987]: I1004 07:11:07.934146 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:07Z","lastTransitionTime":"2025-10-04T07:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.036678 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.036718 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.036727 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.036741 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.036750 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.053661 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.053787 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.138689 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.138733 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.138745 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.138759 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.138770 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.241318 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.241409 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.241443 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.241475 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.241497 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.345007 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.345077 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.345293 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.345317 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.345335 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.448689 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.448859 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.448885 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.448965 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.448988 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.552157 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.552214 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.552232 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.552256 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.552274 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.655026 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.655120 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.655139 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.655171 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.655187 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.758035 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.758108 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.758162 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.758185 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.758198 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.786363 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.786589 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.786567715 +0000 UTC m=+149.841466557 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.860853 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.860890 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.860898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.860913 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.860925 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.888076 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.888166 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.888190 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.888222 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888350 4987 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888397 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888423 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888435 4987 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888398 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888478 4987 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888491 4987 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888565 4987 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888474 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.888445821 +0000 UTC m=+149.943344843 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888741 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.8887175 +0000 UTC m=+149.943616452 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888777 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.888762651 +0000 UTC m=+149.943661703 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 07:11:08 crc kubenswrapper[4987]: E1004 07:11:08.888925 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.888904345 +0000 UTC m=+149.943803367 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.964361 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.964406 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.964416 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.964442 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:08 crc kubenswrapper[4987]: I1004 07:11:08.964457 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:08Z","lastTransitionTime":"2025-10-04T07:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.052813 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.052923 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.052924 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:09 crc kubenswrapper[4987]: E1004 07:11:09.053138 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:09 crc kubenswrapper[4987]: E1004 07:11:09.052953 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:09 crc kubenswrapper[4987]: E1004 07:11:09.053301 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.067895 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.067989 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.068010 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.068037 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.068063 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.170601 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.170735 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.170749 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.170778 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.170793 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.273295 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.273360 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.273376 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.273402 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.273420 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.375860 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.375927 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.375942 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.375969 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.375984 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.478878 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.478926 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.478939 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.478957 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.478968 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.581375 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.581421 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.581431 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.581446 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.581458 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.684202 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.684257 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.684269 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.684291 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.684304 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.788130 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.788183 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.788193 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.788209 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.788219 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.891305 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.891340 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.891349 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.891363 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.891376 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.993956 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.993996 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.994012 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.994028 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:09 crc kubenswrapper[4987]: I1004 07:11:09.994038 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:09Z","lastTransitionTime":"2025-10-04T07:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.053688 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:10 crc kubenswrapper[4987]: E1004 07:11:10.053953 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.096929 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.096984 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.096995 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.097013 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.097024 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.200089 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.200150 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.200163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.200184 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.200198 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.303105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.303152 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.303161 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.303178 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.303188 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.405865 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.406033 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.406104 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.406169 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.406225 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.508420 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.508464 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.508480 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.508505 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.508517 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.611729 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.611777 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.611789 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.611812 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.611824 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.714878 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.714936 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.714952 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.714974 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.714989 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.817562 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.817617 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.817647 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.817668 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.817681 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.920779 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.921064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.921141 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.921214 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:10 crc kubenswrapper[4987]: I1004 07:11:10.921279 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:10Z","lastTransitionTime":"2025-10-04T07:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.023972 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.024035 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.024049 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.024078 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.024093 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.052574 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.052859 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.052678 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:11 crc kubenswrapper[4987]: E1004 07:11:11.053143 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:11 crc kubenswrapper[4987]: E1004 07:11:11.052901 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:11 crc kubenswrapper[4987]: E1004 07:11:11.053326 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.127893 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.127971 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.127997 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.128036 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.128064 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.231422 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.231493 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.231517 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.231545 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.231564 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.335550 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.335705 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.335744 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.335778 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.335800 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.438969 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.439031 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.439040 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.439056 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.439066 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.542962 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.543597 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.543933 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.543965 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.543981 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.647395 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.647439 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.647452 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.647470 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.647484 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.751235 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.751296 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.751307 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.751327 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.751340 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.854373 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.854482 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.854502 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.854532 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.854552 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.958059 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.958164 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.958356 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.958383 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:11 crc kubenswrapper[4987]: I1004 07:11:11.958404 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:11Z","lastTransitionTime":"2025-10-04T07:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.052956 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:12 crc kubenswrapper[4987]: E1004 07:11:12.053188 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.060983 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.061027 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.061040 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.061064 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.061079 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.164021 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.164710 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.164799 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.164867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.164940 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.268773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.269216 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.269452 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.269707 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.269983 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.373788 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.373896 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.373917 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.373949 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.373974 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.478260 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.478340 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.478363 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.478395 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.478417 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.582096 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.582135 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.582144 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.582163 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.582175 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.685472 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.685546 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.685564 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.685593 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.685611 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.789593 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.789695 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.789706 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.789723 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.789734 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.893105 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.893180 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.893206 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.893240 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.893261 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.996581 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.996674 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.996690 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.996718 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:12 crc kubenswrapper[4987]: I1004 07:11:12.996733 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:12Z","lastTransitionTime":"2025-10-04T07:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.053576 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.053663 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:13 crc kubenswrapper[4987]: E1004 07:11:13.053846 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.053594 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:13 crc kubenswrapper[4987]: E1004 07:11:13.054446 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:13 crc kubenswrapper[4987]: E1004 07:11:13.054764 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.100151 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.100204 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.100222 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.100247 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.100265 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.203662 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.203752 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.203772 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.203805 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.203885 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.306820 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.306868 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.306880 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.306896 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.306908 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.410924 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.410997 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.411015 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.411047 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.411066 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.513802 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.513856 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.513866 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.513884 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.513897 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.617655 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.617718 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.617730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.617753 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.617767 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.720822 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.720898 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.720919 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.720946 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.720963 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.825017 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.825533 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.825661 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.825776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.825860 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.929867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.929931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.929941 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.929957 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:13 crc kubenswrapper[4987]: I1004 07:11:13.929969 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:13Z","lastTransitionTime":"2025-10-04T07:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.032566 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.032606 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.032615 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.032647 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.032659 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.052844 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:14 crc kubenswrapper[4987]: E1004 07:11:14.053250 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.075262 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:11:00Z\\\",\\\"message\\\":\\\"5ca3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 07:11:00.942941 7018 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"97419c58-41c7-41d7-a137-a446f0c7eeb3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:11:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k8rdv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tw9rc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.090511 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9a4bbf6f6137760f3404699a8ff6c6812689092e3b0e0a31a036d9730711fea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9befa03e5fd3b5472c53d7f0c31415cba0e6df725df65036aeda0d2c0372afbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.105485 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"470e19ac-b6f8-4818-bda4-2c38c6967b35\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f1d034d6d5563fb58fa132cd1efb1f3063925b20324b1b845db6daa31f2f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7aa1a591ef5d4f06602876bbac76c167ae6cbca4eb313266b4ffb5ef1449bf41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5819e3e4f18789012b1385a8792c3a6d4ccf2536a708308fdd951bfcd5cfee\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bf17bdac5282a17c71bc0441425c256f89d39eeefbcbd6fcb7f61777562458b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.117505 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.130301 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qt4jr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3beef3ad-079d-4f0b-90ed-ff72194c09ca\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://33479ab17ca2cad59093553d9abf41f0014d64950cd48ba599216f79b4870d87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-66tb7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qt4jr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.137266 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.137307 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.137316 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.137329 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.137339 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.147591 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-d76c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9fad7c88-7e3a-4671-8e85-b6e228b2c9c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f5bc486f943849a1d4973ac0c52b59177bf1702d71e10cfb3f78c80fa3de600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://233c0c1cf097b2a7b09d5289b53680308b00ba39d469d2dcf53dae32228ad73a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3a99d1f60a7c143e2caca5ec5ed76feaa8b9e58da1c400dea13b920b60fbb62\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35d09435c112bdb1a765e6df4f934e23c6769d2231551e75c950c7325a462e96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ecb697f6ee50d07ca4ff67c36ba50d80157e4210f5276387b9ccb031290def3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81ad017d5cdc9a16354e02c562270eb09a58f04676177e1ebc9f1c3ec1d29046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a64bf7c74aa82eab46bb31c9fd42849c67bcf488b0fc3b38cff112f2a46494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:10:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xgsnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-d76c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.164178 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b4d7a8a-0a67-4831-9a78-1369771b81db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t456h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:20Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m6jjp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.185924 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e074876-ad6b-422e-b514-4ecb9c07bf02\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef878afa29024e1cc39f9578faa2f1ed7cc05e26dee075e37f96f49653362401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f995c68364bd89259b247aebb51deeebb25ef047488db3d6dc18dc13f78a88f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d2d8164beb29e92633d21e7db8042a2d2764b921442ceb9770bb66837384663\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63b27a3d6d774e085925a4c341618e0b2e090ca57220dfcbde49a5c4150daa06\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f1fd76288616fac029b91b4f26f6bf1ce4889b1695fca7a27f3ef17db724cea\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T07:10:05Z\\\",\\\"message\\\":\\\"le observer\\\\nW1004 07:10:05.167146 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1004 07:10:05.167266 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1004 07:10:05.168364 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3725279554/tls.crt::/tmp/serving-cert-3725279554/tls.key\\\\\\\"\\\\nI1004 07:10:05.681418 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1004 07:10:05.688017 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1004 07:10:05.688041 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1004 07:10:05.688336 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1004 07:10:05.688351 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1004 07:10:05.696207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1004 07:10:05.696241 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1004 07:10:05.696257 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696264 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 07:10:05.696270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 07:10:05.696274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 07:10:05.696278 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 07:10:05.696283 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1004 07:10:05.697915 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3cbc2ae1b82aa65420be93268cb2e9ae01d45cf73660656ce46b153b4df6bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21311283043d46bb7d9cf5039b5f3d21482d92780ef806c426be68dc63a3ab12\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.197104 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b5df3ee-d123-499d-9358-21faf10d2677\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://419b5b4162ac1a6974eba92cc75be83df8bdb354ea5131ecad57c4598cd0a287\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bd1181a0875d40a8902470781227c809914adf34df338c994ecacdd5ce569e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f00013a377af8f941b7066e479e0cfb0dd87527cc87176e58488a71c5885861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e97259a684f1eaf389ce8b8dde1c80bcfe5a583a0689761e5d82d1eb49eba1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.211002 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.223120 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5747e547ccdf44721b2ba476016a34d38fc4e312bf3997757e15fcf5040876f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.239723 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.239766 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.239776 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.239793 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.239806 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.240085 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-mc8jj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"659797bb-8ec1-4ad8-880e-7aec3b068c46\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T07:10:53Z\\\",\\\"message\\\":\\\"2025-10-04T07:10:08+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9\\\\n2025-10-04T07:10:08+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_85b5afa7-21ad-4e4a-822e-59bf4b263fe9 to /host/opt/cni/bin/\\\\n2025-10-04T07:10:08Z [verbose] multus-daemon started\\\\n2025-10-04T07:10:08Z [verbose] Readiness Indicator file check\\\\n2025-10-04T07:10:53Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7n9dt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-multus\"/\"multus-mc8jj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.255885 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a89a1560-c018-4df7-92aa-795638d35d94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8529808f099f446757fe6c27b8c400fa0aaa6b0e8da1a641377e6f72d72f791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phknb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:06Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-dvv6j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.267545 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-fx5cm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1bc8f96-f083-44b7-8b6d-6954d327dc2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59e172fa1b111e01d24bf4428663934f1d3c80a2d11ca0491402adc4a18d846a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6frph\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:08Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-fx5cm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.279916 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b10dd4b8-8472-48a6-a7c8-2cc62180abd1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:09:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1a9adb0d4267aca8ffc9c2a0559f6382c95b1ad2a955b0835f5d6b15eb27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:09:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5449dd5d37978da0f36d2005784c2d0cbcd526e969df576f97b14eee1a272d7f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T07:09:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T07:09:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:09:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.294330 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8738c81b-0365-44fd-bd47-1a1a3b1716d0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fa02a032d592f907d04d303561975bfec616dc018d83de538df37ad692ede4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbb91a6530e3017d37e94421b26791267d758a2a1d5daf55f1b955ab21e7f51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tzlrr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T07:10:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k48xm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.309375 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://030d010992926a3ec3cd94bff172fbc90991b13f313a529472ef3aa80a0da3be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T07:10:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.321121 4987 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T07:10:04Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T07:11:14Z is after 2025-08-24T17:21:41Z" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.342290 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.342355 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.342366 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.342383 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.342410 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.444892 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.444958 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.444972 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.444996 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.445010 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.547488 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.547544 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.547557 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.547578 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.547594 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.650882 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.650979 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.651009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.651044 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.651068 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.754192 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.754268 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.754314 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.754357 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.754408 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.857772 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.857815 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.857825 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.857846 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.857858 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.961653 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.961726 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.961751 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.961780 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:14 crc kubenswrapper[4987]: I1004 07:11:14.961799 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:14Z","lastTransitionTime":"2025-10-04T07:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.053453 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.053935 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.053985 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:15 crc kubenswrapper[4987]: E1004 07:11:15.054117 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:15 crc kubenswrapper[4987]: E1004 07:11:15.054746 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:15 crc kubenswrapper[4987]: E1004 07:11:15.054544 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.065439 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.065501 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.065520 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.065547 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.065564 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.168773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.168881 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.168896 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.168917 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.168932 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.271721 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.271801 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.271845 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.271882 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.271911 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.375585 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.375657 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.375673 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.375694 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.375710 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.478823 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.478865 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.478875 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.478888 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.478900 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.581687 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.581721 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.581730 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.581744 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.581754 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.685760 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.685819 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.685842 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.685870 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.685889 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.789773 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.789836 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.789855 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.789883 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.789905 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.893941 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.894017 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.894036 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.894062 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.894087 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.996931 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.996974 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.996988 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.997009 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:15 crc kubenswrapper[4987]: I1004 07:11:15.997024 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:15Z","lastTransitionTime":"2025-10-04T07:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.054098 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:16 crc kubenswrapper[4987]: E1004 07:11:16.054445 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.100670 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.100739 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.100756 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.100783 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.100804 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.204948 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.205011 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.205023 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.205050 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.205067 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.308332 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.308417 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.308438 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.308468 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.308491 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.411354 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.411394 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.411407 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.411425 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.411438 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.514867 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.514920 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.514937 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.514962 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.514983 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.574809 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.575725 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.576085 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.576264 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.576385 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.629822 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.629893 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.629915 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.629945 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.629966 4987 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T07:11:16Z","lastTransitionTime":"2025-10-04T07:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.653361 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt"] Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.654213 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.656319 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.656706 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.656925 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.657459 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.701452 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-mc8jj" podStartSLOduration=72.701420184 podStartE2EDuration="1m12.701420184s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.701160485 +0000 UTC m=+93.756059377" watchObservedRunningTime="2025-10-04 07:11:16.701420184 +0000 UTC m=+93.756319056" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.723159 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podStartSLOduration=72.723129721 podStartE2EDuration="1m12.723129721s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.723004967 +0000 UTC m=+93.777903819" watchObservedRunningTime="2025-10-04 07:11:16.723129721 +0000 UTC m=+93.778028573" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.757840 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=26.757804336 podStartE2EDuration="26.757804336s" podCreationTimestamp="2025-10-04 07:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.757144465 +0000 UTC m=+93.812043307" watchObservedRunningTime="2025-10-04 07:11:16.757804336 +0000 UTC m=+93.812703218" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.758073 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-fx5cm" podStartSLOduration=72.758065034 podStartE2EDuration="1m12.758065034s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.74268932 +0000 UTC m=+93.797588162" watchObservedRunningTime="2025-10-04 07:11:16.758065034 +0000 UTC m=+93.812963906" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.779313 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.779283636 podStartE2EDuration="1m10.779283636s" podCreationTimestamp="2025-10-04 07:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.779105431 +0000 UTC m=+93.834004273" watchObservedRunningTime="2025-10-04 07:11:16.779283636 +0000 UTC m=+93.834182518" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.784817 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.784888 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.785149 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.785276 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.785345 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.815355 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=39.815328435 podStartE2EDuration="39.815328435s" podCreationTimestamp="2025-10-04 07:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.799785195 +0000 UTC m=+93.854684037" watchObservedRunningTime="2025-10-04 07:11:16.815328435 +0000 UTC m=+93.870227277" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.876161 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k48xm" podStartSLOduration=71.87614131 podStartE2EDuration="1m11.87614131s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:16.845720242 +0000 UTC m=+93.900619084" watchObservedRunningTime="2025-10-04 07:11:16.87614131 +0000 UTC m=+93.931040152" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.886687 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.886724 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.886775 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.886804 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.886823 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.887706 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.887812 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.888421 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.893281 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.905386 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2qmvt\" (UID: \"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: I1004 07:11:16.976000 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" Oct 04 07:11:16 crc kubenswrapper[4987]: W1004 07:11:16.995322 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cbc3c2e_12c2_4cb5_b825_a9d4932a3f36.slice/crio-56b160838c57c69c092b1cb0e59f88834df6c481256e977cbe0a9f3f97272e39 WatchSource:0}: Error finding container 56b160838c57c69c092b1cb0e59f88834df6c481256e977cbe0a9f3f97272e39: Status 404 returned error can't find the container with id 56b160838c57c69c092b1cb0e59f88834df6c481256e977cbe0a9f3f97272e39 Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.002451 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-d76c6" podStartSLOduration=73.0024334 podStartE2EDuration="1m13.0024334s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:17.002048107 +0000 UTC m=+94.056946949" watchObservedRunningTime="2025-10-04 07:11:17.0024334 +0000 UTC m=+94.057332242" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.045642 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=73.045606637 podStartE2EDuration="1m13.045606637s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:17.044604786 +0000 UTC m=+94.099503628" watchObservedRunningTime="2025-10-04 07:11:17.045606637 +0000 UTC m=+94.100505479" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.053117 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.053182 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:17 crc kubenswrapper[4987]: E1004 07:11:17.053247 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:17 crc kubenswrapper[4987]: E1004 07:11:17.053289 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.053526 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:17 crc kubenswrapper[4987]: E1004 07:11:17.053595 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.054111 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:11:17 crc kubenswrapper[4987]: E1004 07:11:17.054347 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.072833 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qt4jr" podStartSLOduration=73.072806422 podStartE2EDuration="1m13.072806422s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:17.071949034 +0000 UTC m=+94.126847896" watchObservedRunningTime="2025-10-04 07:11:17.072806422 +0000 UTC m=+94.127705264" Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.528737 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" event={"ID":"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36","Type":"ContainerStarted","Data":"63727aff6479d6a91e807a000417e9416b1cc7a552a23e01d7418fc68a78f328"} Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.528810 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" event={"ID":"0cbc3c2e-12c2-4cb5-b825-a9d4932a3f36","Type":"ContainerStarted","Data":"56b160838c57c69c092b1cb0e59f88834df6c481256e977cbe0a9f3f97272e39"} Oct 04 07:11:17 crc kubenswrapper[4987]: I1004 07:11:17.546496 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2qmvt" podStartSLOduration=73.546473248 podStartE2EDuration="1m13.546473248s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:17.545271309 +0000 UTC m=+94.600170151" watchObservedRunningTime="2025-10-04 07:11:17.546473248 +0000 UTC m=+94.601372110" Oct 04 07:11:18 crc kubenswrapper[4987]: I1004 07:11:18.053121 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:18 crc kubenswrapper[4987]: E1004 07:11:18.053244 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:19 crc kubenswrapper[4987]: I1004 07:11:19.053165 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:19 crc kubenswrapper[4987]: I1004 07:11:19.053275 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:19 crc kubenswrapper[4987]: E1004 07:11:19.053329 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:19 crc kubenswrapper[4987]: I1004 07:11:19.053442 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:19 crc kubenswrapper[4987]: E1004 07:11:19.053525 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:19 crc kubenswrapper[4987]: E1004 07:11:19.053788 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:20 crc kubenswrapper[4987]: I1004 07:11:20.052619 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:20 crc kubenswrapper[4987]: E1004 07:11:20.052792 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:21 crc kubenswrapper[4987]: I1004 07:11:21.052710 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:21 crc kubenswrapper[4987]: I1004 07:11:21.052746 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:21 crc kubenswrapper[4987]: I1004 07:11:21.052883 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:21 crc kubenswrapper[4987]: E1004 07:11:21.053144 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:21 crc kubenswrapper[4987]: E1004 07:11:21.054002 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:21 crc kubenswrapper[4987]: E1004 07:11:21.054076 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:21 crc kubenswrapper[4987]: I1004 07:11:21.070446 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 07:11:22 crc kubenswrapper[4987]: I1004 07:11:22.053049 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:22 crc kubenswrapper[4987]: E1004 07:11:22.053246 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:23 crc kubenswrapper[4987]: I1004 07:11:23.052496 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:23 crc kubenswrapper[4987]: I1004 07:11:23.052537 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:23 crc kubenswrapper[4987]: I1004 07:11:23.052553 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:23 crc kubenswrapper[4987]: E1004 07:11:23.053008 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:23 crc kubenswrapper[4987]: E1004 07:11:23.053138 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:23 crc kubenswrapper[4987]: E1004 07:11:23.053203 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:24 crc kubenswrapper[4987]: I1004 07:11:24.052578 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:24 crc kubenswrapper[4987]: E1004 07:11:24.054791 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:24 crc kubenswrapper[4987]: I1004 07:11:24.093340 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.093307389 podStartE2EDuration="3.093307389s" podCreationTimestamp="2025-10-04 07:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:24.091341756 +0000 UTC m=+101.146240668" watchObservedRunningTime="2025-10-04 07:11:24.093307389 +0000 UTC m=+101.148206271" Oct 04 07:11:24 crc kubenswrapper[4987]: I1004 07:11:24.373215 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:24 crc kubenswrapper[4987]: E1004 07:11:24.373408 4987 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:11:24 crc kubenswrapper[4987]: E1004 07:11:24.373490 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs podName:4b4d7a8a-0a67-4831-9a78-1369771b81db nodeName:}" failed. No retries permitted until 2025-10-04 07:12:28.373468816 +0000 UTC m=+165.428367668 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs") pod "network-metrics-daemon-m6jjp" (UID: "4b4d7a8a-0a67-4831-9a78-1369771b81db") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 07:11:25 crc kubenswrapper[4987]: I1004 07:11:25.052579 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:25 crc kubenswrapper[4987]: I1004 07:11:25.053013 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:25 crc kubenswrapper[4987]: E1004 07:11:25.053179 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:25 crc kubenswrapper[4987]: I1004 07:11:25.053233 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:25 crc kubenswrapper[4987]: E1004 07:11:25.053244 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:25 crc kubenswrapper[4987]: E1004 07:11:25.053367 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:26 crc kubenswrapper[4987]: I1004 07:11:26.053829 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:26 crc kubenswrapper[4987]: E1004 07:11:26.053940 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:27 crc kubenswrapper[4987]: I1004 07:11:27.053360 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:27 crc kubenswrapper[4987]: I1004 07:11:27.053399 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:27 crc kubenswrapper[4987]: E1004 07:11:27.053500 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:27 crc kubenswrapper[4987]: I1004 07:11:27.053436 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:27 crc kubenswrapper[4987]: E1004 07:11:27.053619 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:27 crc kubenswrapper[4987]: E1004 07:11:27.053794 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:28 crc kubenswrapper[4987]: I1004 07:11:28.053941 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:28 crc kubenswrapper[4987]: E1004 07:11:28.054107 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:29 crc kubenswrapper[4987]: I1004 07:11:29.053170 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:29 crc kubenswrapper[4987]: I1004 07:11:29.053234 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:29 crc kubenswrapper[4987]: I1004 07:11:29.053203 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:29 crc kubenswrapper[4987]: E1004 07:11:29.053344 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:29 crc kubenswrapper[4987]: E1004 07:11:29.053470 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:29 crc kubenswrapper[4987]: E1004 07:11:29.053558 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:30 crc kubenswrapper[4987]: I1004 07:11:30.053265 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:30 crc kubenswrapper[4987]: E1004 07:11:30.053430 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:31 crc kubenswrapper[4987]: I1004 07:11:31.053440 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:31 crc kubenswrapper[4987]: I1004 07:11:31.053473 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:31 crc kubenswrapper[4987]: E1004 07:11:31.053562 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:31 crc kubenswrapper[4987]: E1004 07:11:31.053849 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:31 crc kubenswrapper[4987]: I1004 07:11:31.054944 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:31 crc kubenswrapper[4987]: E1004 07:11:31.055881 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:31 crc kubenswrapper[4987]: I1004 07:11:31.056429 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:11:31 crc kubenswrapper[4987]: E1004 07:11:31.056857 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tw9rc_openshift-ovn-kubernetes(c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" Oct 04 07:11:32 crc kubenswrapper[4987]: I1004 07:11:32.052953 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:32 crc kubenswrapper[4987]: E1004 07:11:32.053131 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:33 crc kubenswrapper[4987]: I1004 07:11:33.053868 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:33 crc kubenswrapper[4987]: I1004 07:11:33.054026 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:33 crc kubenswrapper[4987]: E1004 07:11:33.054109 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:33 crc kubenswrapper[4987]: E1004 07:11:33.054213 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:33 crc kubenswrapper[4987]: I1004 07:11:33.054338 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:33 crc kubenswrapper[4987]: E1004 07:11:33.054550 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:34 crc kubenswrapper[4987]: I1004 07:11:34.053295 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:34 crc kubenswrapper[4987]: E1004 07:11:34.054043 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:35 crc kubenswrapper[4987]: I1004 07:11:35.053525 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:35 crc kubenswrapper[4987]: I1004 07:11:35.053550 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:35 crc kubenswrapper[4987]: I1004 07:11:35.053581 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:35 crc kubenswrapper[4987]: E1004 07:11:35.053885 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:35 crc kubenswrapper[4987]: E1004 07:11:35.054017 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:35 crc kubenswrapper[4987]: E1004 07:11:35.054107 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:36 crc kubenswrapper[4987]: I1004 07:11:36.053324 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:36 crc kubenswrapper[4987]: E1004 07:11:36.053789 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:37 crc kubenswrapper[4987]: I1004 07:11:37.052889 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:37 crc kubenswrapper[4987]: I1004 07:11:37.053096 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:37 crc kubenswrapper[4987]: E1004 07:11:37.053286 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:37 crc kubenswrapper[4987]: I1004 07:11:37.053405 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:37 crc kubenswrapper[4987]: E1004 07:11:37.053575 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:37 crc kubenswrapper[4987]: E1004 07:11:37.054127 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:38 crc kubenswrapper[4987]: I1004 07:11:38.053362 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:38 crc kubenswrapper[4987]: E1004 07:11:38.053494 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:39 crc kubenswrapper[4987]: I1004 07:11:39.053268 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:39 crc kubenswrapper[4987]: I1004 07:11:39.053343 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:39 crc kubenswrapper[4987]: I1004 07:11:39.053372 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:39 crc kubenswrapper[4987]: E1004 07:11:39.053413 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:39 crc kubenswrapper[4987]: E1004 07:11:39.053471 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:39 crc kubenswrapper[4987]: E1004 07:11:39.053547 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.052828 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:40 crc kubenswrapper[4987]: E1004 07:11:40.053043 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.604333 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/1.log" Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.605406 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/0.log" Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.605480 4987 generic.go:334] "Generic (PLEG): container finished" podID="659797bb-8ec1-4ad8-880e-7aec3b068c46" containerID="d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942" exitCode=1 Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.605516 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerDied","Data":"d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942"} Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.605557 4987 scope.go:117] "RemoveContainer" containerID="d9596a1f3943577cde311044239f2147ca71d6f987dcf006ef5f0a94c9e11599" Oct 04 07:11:40 crc kubenswrapper[4987]: I1004 07:11:40.606123 4987 scope.go:117] "RemoveContainer" containerID="d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942" Oct 04 07:11:40 crc kubenswrapper[4987]: E1004 07:11:40.606332 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-mc8jj_openshift-multus(659797bb-8ec1-4ad8-880e-7aec3b068c46)\"" pod="openshift-multus/multus-mc8jj" podUID="659797bb-8ec1-4ad8-880e-7aec3b068c46" Oct 04 07:11:41 crc kubenswrapper[4987]: I1004 07:11:41.052486 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:41 crc kubenswrapper[4987]: I1004 07:11:41.052572 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:41 crc kubenswrapper[4987]: E1004 07:11:41.052616 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:41 crc kubenswrapper[4987]: I1004 07:11:41.052749 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:41 crc kubenswrapper[4987]: E1004 07:11:41.052899 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:41 crc kubenswrapper[4987]: E1004 07:11:41.053071 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:41 crc kubenswrapper[4987]: I1004 07:11:41.610964 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/1.log" Oct 04 07:11:42 crc kubenswrapper[4987]: I1004 07:11:42.053003 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:42 crc kubenswrapper[4987]: E1004 07:11:42.053217 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:43 crc kubenswrapper[4987]: I1004 07:11:43.052440 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:43 crc kubenswrapper[4987]: I1004 07:11:43.052463 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:43 crc kubenswrapper[4987]: E1004 07:11:43.052557 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:43 crc kubenswrapper[4987]: I1004 07:11:43.052588 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:43 crc kubenswrapper[4987]: E1004 07:11:43.052786 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:43 crc kubenswrapper[4987]: E1004 07:11:43.052873 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:44 crc kubenswrapper[4987]: E1004 07:11:44.048905 4987 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 07:11:44 crc kubenswrapper[4987]: I1004 07:11:44.053204 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:44 crc kubenswrapper[4987]: E1004 07:11:44.060000 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:44 crc kubenswrapper[4987]: E1004 07:11:44.145806 4987 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 07:11:45 crc kubenswrapper[4987]: I1004 07:11:45.052883 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:45 crc kubenswrapper[4987]: E1004 07:11:45.053027 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:45 crc kubenswrapper[4987]: I1004 07:11:45.053057 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:45 crc kubenswrapper[4987]: I1004 07:11:45.053189 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:45 crc kubenswrapper[4987]: E1004 07:11:45.053483 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:45 crc kubenswrapper[4987]: E1004 07:11:45.053799 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.053454 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:46 crc kubenswrapper[4987]: E1004 07:11:46.053738 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.054374 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.629839 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/3.log" Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.632915 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerStarted","Data":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.633316 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.669452 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podStartSLOduration=101.669426295 podStartE2EDuration="1m41.669426295s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:11:46.668612289 +0000 UTC m=+123.723511141" watchObservedRunningTime="2025-10-04 07:11:46.669426295 +0000 UTC m=+123.724325137" Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.910094 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m6jjp"] Oct 04 07:11:46 crc kubenswrapper[4987]: I1004 07:11:46.910204 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:46 crc kubenswrapper[4987]: E1004 07:11:46.910290 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:47 crc kubenswrapper[4987]: I1004 07:11:47.053190 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:47 crc kubenswrapper[4987]: I1004 07:11:47.053199 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:47 crc kubenswrapper[4987]: I1004 07:11:47.053327 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:47 crc kubenswrapper[4987]: E1004 07:11:47.053466 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:47 crc kubenswrapper[4987]: E1004 07:11:47.053695 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:47 crc kubenswrapper[4987]: E1004 07:11:47.053861 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:49 crc kubenswrapper[4987]: I1004 07:11:49.052818 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:49 crc kubenswrapper[4987]: E1004 07:11:49.053250 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:49 crc kubenswrapper[4987]: I1004 07:11:49.053473 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:49 crc kubenswrapper[4987]: I1004 07:11:49.053792 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:49 crc kubenswrapper[4987]: E1004 07:11:49.053919 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:49 crc kubenswrapper[4987]: I1004 07:11:49.053792 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:49 crc kubenswrapper[4987]: E1004 07:11:49.053986 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:49 crc kubenswrapper[4987]: E1004 07:11:49.054004 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:49 crc kubenswrapper[4987]: E1004 07:11:49.146840 4987 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 07:11:51 crc kubenswrapper[4987]: I1004 07:11:51.053399 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:51 crc kubenswrapper[4987]: E1004 07:11:51.053540 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:51 crc kubenswrapper[4987]: I1004 07:11:51.053785 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:51 crc kubenswrapper[4987]: I1004 07:11:51.053784 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:51 crc kubenswrapper[4987]: I1004 07:11:51.053870 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:51 crc kubenswrapper[4987]: E1004 07:11:51.054159 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:51 crc kubenswrapper[4987]: E1004 07:11:51.054275 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:51 crc kubenswrapper[4987]: E1004 07:11:51.054333 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:53 crc kubenswrapper[4987]: I1004 07:11:53.053129 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:53 crc kubenswrapper[4987]: E1004 07:11:53.053254 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:53 crc kubenswrapper[4987]: I1004 07:11:53.054650 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:53 crc kubenswrapper[4987]: E1004 07:11:53.054712 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:53 crc kubenswrapper[4987]: I1004 07:11:53.054825 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:53 crc kubenswrapper[4987]: E1004 07:11:53.054985 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:53 crc kubenswrapper[4987]: I1004 07:11:53.054660 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:53 crc kubenswrapper[4987]: E1004 07:11:53.055587 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:54 crc kubenswrapper[4987]: E1004 07:11:54.147364 4987 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.053594 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.053668 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.053687 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:55 crc kubenswrapper[4987]: E1004 07:11:55.053926 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:55 crc kubenswrapper[4987]: E1004 07:11:55.054062 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:55 crc kubenswrapper[4987]: E1004 07:11:55.054152 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.054233 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.054254 4987 scope.go:117] "RemoveContainer" containerID="d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942" Oct 04 07:11:55 crc kubenswrapper[4987]: E1004 07:11:55.054434 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.665245 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/1.log" Oct 04 07:11:55 crc kubenswrapper[4987]: I1004 07:11:55.665289 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerStarted","Data":"1db87dc841e81f4e40e9a6c80e98c312d2b1ab807484d90d9f2e1be091bf01ad"} Oct 04 07:11:57 crc kubenswrapper[4987]: I1004 07:11:57.053063 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:57 crc kubenswrapper[4987]: I1004 07:11:57.053123 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:57 crc kubenswrapper[4987]: I1004 07:11:57.053143 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:57 crc kubenswrapper[4987]: E1004 07:11:57.053982 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:57 crc kubenswrapper[4987]: E1004 07:11:57.053722 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:57 crc kubenswrapper[4987]: I1004 07:11:57.053162 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:57 crc kubenswrapper[4987]: E1004 07:11:57.054078 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:11:57 crc kubenswrapper[4987]: E1004 07:11:57.054197 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:59 crc kubenswrapper[4987]: I1004 07:11:59.053693 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:11:59 crc kubenswrapper[4987]: I1004 07:11:59.053852 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:11:59 crc kubenswrapper[4987]: I1004 07:11:59.053935 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:11:59 crc kubenswrapper[4987]: I1004 07:11:59.053860 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:11:59 crc kubenswrapper[4987]: E1004 07:11:59.054091 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m6jjp" podUID="4b4d7a8a-0a67-4831-9a78-1369771b81db" Oct 04 07:11:59 crc kubenswrapper[4987]: E1004 07:11:59.054237 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 07:11:59 crc kubenswrapper[4987]: E1004 07:11:59.054417 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 07:11:59 crc kubenswrapper[4987]: E1004 07:11:59.054537 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.053165 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.053182 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.053353 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.053354 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.055297 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.055748 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.055803 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.055969 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.056145 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 07:12:01 crc kubenswrapper[4987]: I1004 07:12:01.056147 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.689170 4987 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.732771 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tfd59"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.734130 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.736588 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wt287"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.737600 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.737952 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.738091 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.738265 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.739151 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.738325 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.738459 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.738523 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.740308 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.743579 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.758862 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.762513 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.763243 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.763771 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.764040 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.764743 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.764804 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.765086 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.761211 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.765798 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.761746 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.762048 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.762218 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.762241 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.766712 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.784114 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.786878 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.787129 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.787357 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.791824 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w4dpg"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.792432 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.793035 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.793051 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.793258 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.793283 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.793288 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.795992 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.804198 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f49kw"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.804812 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.804984 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.805292 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.805489 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.805551 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.805898 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806013 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806109 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806219 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806221 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806306 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806333 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806391 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806461 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806551 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806574 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806813 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.806929 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wfp5w"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.807304 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.808054 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6hddn"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.808728 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.808889 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.809336 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.812895 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rgcdb"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.813263 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-t2w8v"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.813610 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bbx6q"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.814152 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.814676 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.814925 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.815244 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.815308 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mrxhg"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.815615 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.815861 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.816180 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.818284 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.818463 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.818659 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.818942 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.819253 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.819513 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.829694 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.832390 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.833277 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.833456 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.833832 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.834015 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.833875 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.821105 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.820933 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.821282 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.821307 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.834298 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.834533 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.834831 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.834992 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.821245 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.835425 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.836191 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.836952 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.844317 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.838729 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.837353 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.837338 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.837503 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.843982 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.844029 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.846395 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.846979 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.847022 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.847139 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.847238 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.847380 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.850576 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.850854 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.851371 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.852642 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.853831 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.853991 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.853988 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.854058 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.854178 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.854300 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.854920 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbp5h\" (UniqueName: \"kubernetes.io/projected/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-kube-api-access-vbp5h\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.855020 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrbhp\" (UniqueName: \"kubernetes.io/projected/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-kube-api-access-jrbhp\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.855080 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-audit-policies\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.855113 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-encryption-config\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.855144 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wllzd\" (UniqueName: \"kubernetes.io/projected/1a5130dd-df25-4642-8180-637b14f8ee4a-kube-api-access-wllzd\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.855175 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-etcd-serving-ca\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.867665 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.868078 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5b448"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.868472 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.868876 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.869382 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.869787 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870078 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870222 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870353 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870539 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-etcd-client\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870704 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-config\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870806 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzk5b\" (UniqueName: \"kubernetes.io/projected/262ca319-a044-47bb-b107-1b76e9d9abbb-kube-api-access-lzk5b\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870923 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-config\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871127 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-config\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871259 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-etcd-client\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871369 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871469 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871573 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-serving-cert\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871692 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.870603 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871891 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-audit\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871987 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-encryption-config\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.872100 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-image-import-ca\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871585 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.871792 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.872665 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-audit-dir\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.872777 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.872873 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-client-ca\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.872981 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a5130dd-df25-4642-8180-637b14f8ee4a-serving-cert\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.873075 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/262ca319-a044-47bb-b107-1b76e9d9abbb-node-pullsecrets\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.873176 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/262ca319-a044-47bb-b107-1b76e9d9abbb-audit-dir\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.873352 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-auth-proxy-config\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.873527 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-serving-cert\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.873674 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-machine-approver-tls\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.876862 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.878149 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bjgbf"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.879203 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.883126 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.883802 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.886728 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.887393 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.887478 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.888409 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.893843 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.894878 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.899336 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.899916 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.903041 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.903296 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.903816 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.906561 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.907146 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.907452 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.910932 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qv6pn"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.911204 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.911484 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.912181 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.913044 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.920030 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.920738 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.921203 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.921672 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.923558 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.924587 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.924712 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.926049 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.928231 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.930017 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.932302 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.935327 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.935971 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.936914 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9wm6"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.937998 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.940515 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tfd59"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.952592 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zzpsj"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.953600 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.955171 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.955895 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.956746 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.957317 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.957661 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.958542 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4786l"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.959452 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.960921 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.966005 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.966730 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.968662 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.969573 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.970518 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wt287"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.972745 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mrxhg"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.973246 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w4dpg"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.974357 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.975448 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.976205 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-audit-policies\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977067 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wllzd\" (UniqueName: \"kubernetes.io/projected/1a5130dd-df25-4642-8180-637b14f8ee4a-kube-api-access-wllzd\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977099 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-etcd-serving-ca\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977124 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-encryption-config\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977144 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-etcd-client\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977164 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b67efca-532b-437c-b377-2c3bfb7372c0-metrics-tls\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977032 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-audit-policies\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.976848 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bbx6q"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977184 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr5v4\" (UniqueName: \"kubernetes.io/projected/3b67efca-532b-437c-b377-2c3bfb7372c0-kube-api-access-nr5v4\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977293 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-service-ca-bundle\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977173 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977390 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-config\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977416 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-config\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977438 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzk5b\" (UniqueName: \"kubernetes.io/projected/262ca319-a044-47bb-b107-1b76e9d9abbb-kube-api-access-lzk5b\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.977981 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-etcd-serving-ca\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978067 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-config\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978332 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-config\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978444 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-config\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978498 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-etcd-client\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978528 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978547 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978569 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978586 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-serving-cert\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978612 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978648 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdppg\" (UniqueName: \"kubernetes.io/projected/453b9a94-8a64-477b-97eb-633594674f79-kube-api-access-kdppg\") pod \"downloads-7954f5f757-t2w8v\" (UID: \"453b9a94-8a64-477b-97eb-633594674f79\") " pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978672 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-audit\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.978688 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-encryption-config\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979008 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9f070f-4908-41d2-b247-f0805b1c4792-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979043 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-image-import-ca\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979064 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nd56\" (UniqueName: \"kubernetes.io/projected/e435b9eb-8082-4935-8d71-14ea61d5119a-kube-api-access-6nd56\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979091 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-audit-dir\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979111 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a5130dd-df25-4642-8180-637b14f8ee4a-serving-cert\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979131 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/262ca319-a044-47bb-b107-1b76e9d9abbb-node-pullsecrets\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979150 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/262ca319-a044-47bb-b107-1b76e9d9abbb-audit-dir\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979178 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979196 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-client-ca\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979222 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-auth-proxy-config\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979254 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-config\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979306 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-audit-dir\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.979350 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b67efca-532b-437c-b377-2c3bfb7372c0-trusted-ca\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.980454 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.980521 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.980580 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/262ca319-a044-47bb-b107-1b76e9d9abbb-audit-dir\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.980593 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-audit\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.981068 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/262ca319-a044-47bb-b107-1b76e9d9abbb-node-pullsecrets\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.981796 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.981923 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t547r\" (UniqueName: \"kubernetes.io/projected/fb9f070f-4908-41d2-b247-f0805b1c4792-kube-api-access-t547r\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.981964 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqtld\" (UniqueName: \"kubernetes.io/projected/acc7a457-28b0-4bd8-98aa-85110a95ae02-kube-api-access-sqtld\") pod \"migrator-59844c95c7-th4rv\" (UID: \"acc7a457-28b0-4bd8-98aa-85110a95ae02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982042 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-serving-cert\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982098 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-machine-approver-tls\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982142 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e435b9eb-8082-4935-8d71-14ea61d5119a-serving-cert\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982412 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982478 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b67efca-532b-437c-b377-2c3bfb7372c0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982526 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb9f070f-4908-41d2-b247-f0805b1c4792-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982584 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbp5h\" (UniqueName: \"kubernetes.io/projected/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-kube-api-access-vbp5h\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982646 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrbhp\" (UniqueName: \"kubernetes.io/projected/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-kube-api-access-jrbhp\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.982934 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-etcd-client\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.983311 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.983650 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-etcd-client\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.984248 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-encryption-config\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.985205 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-client-ca\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.985317 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.985402 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/262ca319-a044-47bb-b107-1b76e9d9abbb-image-import-ca\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.985406 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a5130dd-df25-4642-8180-637b14f8ee4a-serving-cert\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.986779 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/262ca319-a044-47bb-b107-1b76e9d9abbb-serving-cert\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.987686 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.987964 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-auth-proxy-config\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.988097 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wfp5w"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.989219 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6hddn"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.989232 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-config\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.990088 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-serving-cert\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.990731 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-t2w8v"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.991887 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-machine-approver-tls\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.992117 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.994885 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.994922 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.995352 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qv6pn"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.995552 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-encryption-config\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.996607 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rgcdb"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.997227 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.997867 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vzg8c"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.998951 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f"] Oct 04 07:12:07 crc kubenswrapper[4987]: I1004 07:12:07.999198 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.000259 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.001394 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.002504 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.005006 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.006790 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f49kw"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.008945 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5b448"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.012317 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.013765 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.015487 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zzpsj"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.017220 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.017509 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.018619 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.021265 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.030752 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.032026 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.033905 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.036072 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9wm6"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.038440 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.040478 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4786l"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.042112 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pdb9r"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.043674 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-nv9l8"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.044012 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.044313 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.045246 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pdb9r"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.046692 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nv9l8"] Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.057415 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.076995 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083696 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9f070f-4908-41d2-b247-f0805b1c4792-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083735 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nd56\" (UniqueName: \"kubernetes.io/projected/e435b9eb-8082-4935-8d71-14ea61d5119a-kube-api-access-6nd56\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083779 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-config\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083802 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b67efca-532b-437c-b377-2c3bfb7372c0-trusted-ca\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083821 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t547r\" (UniqueName: \"kubernetes.io/projected/fb9f070f-4908-41d2-b247-f0805b1c4792-kube-api-access-t547r\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083841 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqtld\" (UniqueName: \"kubernetes.io/projected/acc7a457-28b0-4bd8-98aa-85110a95ae02-kube-api-access-sqtld\") pod \"migrator-59844c95c7-th4rv\" (UID: \"acc7a457-28b0-4bd8-98aa-85110a95ae02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083860 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e435b9eb-8082-4935-8d71-14ea61d5119a-serving-cert\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083878 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b67efca-532b-437c-b377-2c3bfb7372c0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083895 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb9f070f-4908-41d2-b247-f0805b1c4792-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083949 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b67efca-532b-437c-b377-2c3bfb7372c0-metrics-tls\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.083979 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr5v4\" (UniqueName: \"kubernetes.io/projected/3b67efca-532b-437c-b377-2c3bfb7372c0-kube-api-access-nr5v4\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.084007 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-service-ca-bundle\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.084050 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.084071 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdppg\" (UniqueName: \"kubernetes.io/projected/453b9a94-8a64-477b-97eb-633594674f79-kube-api-access-kdppg\") pod \"downloads-7954f5f757-t2w8v\" (UID: \"453b9a94-8a64-477b-97eb-633594674f79\") " pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.084826 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-config\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.084850 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9f070f-4908-41d2-b247-f0805b1c4792-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.084908 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-service-ca-bundle\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.085219 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e435b9eb-8082-4935-8d71-14ea61d5119a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.085463 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b67efca-532b-437c-b377-2c3bfb7372c0-trusted-ca\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.087381 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e435b9eb-8082-4935-8d71-14ea61d5119a-serving-cert\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.087396 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b67efca-532b-437c-b377-2c3bfb7372c0-metrics-tls\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.087905 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb9f070f-4908-41d2-b247-f0805b1c4792-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.097460 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.118199 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.137644 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.158145 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.177094 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.197533 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.217375 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.238763 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.257821 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.277207 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.297336 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.317499 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.337571 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.356679 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.376919 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.397453 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.416421 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.437821 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.458602 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.477854 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.497598 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.517323 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.537789 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.557448 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.577807 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.597443 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.616923 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.638153 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.657003 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.697687 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.717053 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.737280 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.757455 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.777063 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.797577 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.817006 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.837952 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.857872 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.877245 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.915683 4987 request.go:700] Waited for 1.002274616s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-ac-dockercfg-9lkdf&limit=500&resourceVersion=0 Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.918308 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.939123 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.957529 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.977923 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 07:12:08 crc kubenswrapper[4987]: I1004 07:12:08.997805 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.018267 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.037892 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.058578 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.076657 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.096969 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.116709 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.136261 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.163964 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.176875 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.196829 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.217227 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.236487 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.256550 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.277006 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.296894 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.317717 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.337385 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.357110 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.376989 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.398380 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.417719 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.437776 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.457085 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.477195 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.497318 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.516762 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.537567 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.558036 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.577431 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.597460 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.617605 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.655492 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wllzd\" (UniqueName: \"kubernetes.io/projected/1a5130dd-df25-4642-8180-637b14f8ee4a-kube-api-access-wllzd\") pod \"controller-manager-879f6c89f-tfd59\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.689921 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzk5b\" (UniqueName: \"kubernetes.io/projected/262ca319-a044-47bb-b107-1b76e9d9abbb-kube-api-access-lzk5b\") pod \"apiserver-76f77b778f-wt287\" (UID: \"262ca319-a044-47bb-b107-1b76e9d9abbb\") " pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.701219 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbp5h\" (UniqueName: \"kubernetes.io/projected/25e66a14-b32b-4aca-9fce-9fc81f5d66f4-kube-api-access-vbp5h\") pod \"machine-approver-56656f9798-rnb5t\" (UID: \"25e66a14-b32b-4aca-9fce-9fc81f5d66f4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.711719 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrbhp\" (UniqueName: \"kubernetes.io/projected/8e6f14f4-f9b0-438a-a4b4-290b3beb7814-kube-api-access-jrbhp\") pod \"apiserver-7bbb656c7d-wxvx4\" (UID: \"8e6f14f4-f9b0-438a-a4b4-290b3beb7814\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.717055 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.737172 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.757004 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.777337 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.797986 4987 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.817946 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.837294 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.858226 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.876651 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.899272 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.912057 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.922752 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nd56\" (UniqueName: \"kubernetes.io/projected/e435b9eb-8082-4935-8d71-14ea61d5119a-kube-api-access-6nd56\") pod \"authentication-operator-69f744f599-w4dpg\" (UID: \"e435b9eb-8082-4935-8d71-14ea61d5119a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.930765 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqtld\" (UniqueName: \"kubernetes.io/projected/acc7a457-28b0-4bd8-98aa-85110a95ae02-kube-api-access-sqtld\") pod \"migrator-59844c95c7-th4rv\" (UID: \"acc7a457-28b0-4bd8-98aa-85110a95ae02\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.934877 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.935404 4987 request.go:700] Waited for 1.851325746s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager-operator/serviceaccounts/openshift-controller-manager-operator/token Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.964432 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t547r\" (UniqueName: \"kubernetes.io/projected/fb9f070f-4908-41d2-b247-f0805b1c4792-kube-api-access-t547r\") pod \"openshift-controller-manager-operator-756b6f6bc6-9mxgt\" (UID: \"fb9f070f-4908-41d2-b247-f0805b1c4792\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.976040 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b67efca-532b-437c-b377-2c3bfb7372c0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.978275 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:09 crc kubenswrapper[4987]: I1004 07:12:09.985217 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.000925 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr5v4\" (UniqueName: \"kubernetes.io/projected/3b67efca-532b-437c-b377-2c3bfb7372c0-kube-api-access-nr5v4\") pod \"ingress-operator-5b745b69d9-7hkqx\" (UID: \"3b67efca-532b-437c-b377-2c3bfb7372c0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.010596 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.016431 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdppg\" (UniqueName: \"kubernetes.io/projected/453b9a94-8a64-477b-97eb-633594674f79-kube-api-access-kdppg\") pod \"downloads-7954f5f757-t2w8v\" (UID: \"453b9a94-8a64-477b-97eb-633594674f79\") " pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105265 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-config\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105300 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105318 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105358 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3206d6a-5f6e-4f19-9dac-93b2689734dc-serving-cert\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105474 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ba6778c-5684-46a3-b7c5-75f26973e3e0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105515 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2b4v\" (UniqueName: \"kubernetes.io/projected/9021090f-c067-4574-979a-4186a2df4b89-kube-api-access-p2b4v\") pod \"cluster-samples-operator-665b6dd947-96m8f\" (UID: \"9021090f-c067-4574-979a-4186a2df4b89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105538 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105559 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105579 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-policies\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105595 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105641 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/001408b5-e454-432d-93dc-3a2dc497941f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105657 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmx8b\" (UniqueName: \"kubernetes.io/projected/d5787789-e288-4011-baeb-17da6439f196-kube-api-access-jmx8b\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105672 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-default-certificate\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105693 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4vwr\" (UniqueName: \"kubernetes.io/projected/dc8c0c59-a185-4b91-920e-c53f58d27e61-kube-api-access-l4vwr\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105717 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-client\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105760 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5787789-e288-4011-baeb-17da6439f196-images\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105782 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ba6778c-5684-46a3-b7c5-75f26973e3e0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105802 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.105953 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm5rb\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-kube-api-access-dm5rb\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106115 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-stats-auth\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106165 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggmhj\" (UniqueName: \"kubernetes.io/projected/5420a718-1d0a-4b72-b57d-a8a4a18e7763-kube-api-access-ggmhj\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdbxg\" (UID: \"5420a718-1d0a-4b72-b57d-a8a4a18e7763\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106186 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e1f11da-af03-4bc8-97c8-ad2eac21d489-service-ca-bundle\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106243 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106304 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbsz5\" (UniqueName: \"kubernetes.io/projected/e3e5bf21-b590-45a3-9891-3e5114d6b25d-kube-api-access-xbsz5\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106409 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddmnb\" (UniqueName: \"kubernetes.io/projected/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-kube-api-access-ddmnb\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106447 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-client-ca\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106493 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-oauth-config\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106534 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpd47\" (UniqueName: \"kubernetes.io/projected/e2745959-fc00-405c-8c5d-d70e7cb132ec-kube-api-access-zpd47\") pod \"dns-operator-744455d44c-bbx6q\" (UID: \"e2745959-fc00-405c-8c5d-d70e7cb132ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106554 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-serving-cert\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106591 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-service-ca\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106609 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106653 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5420a718-1d0a-4b72-b57d-a8a4a18e7763-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdbxg\" (UID: \"5420a718-1d0a-4b72-b57d-a8a4a18e7763\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106689 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-trusted-ca\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106707 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2745959-fc00-405c-8c5d-d70e7cb132ec-metrics-tls\") pod \"dns-operator-744455d44c-bbx6q\" (UID: \"e2745959-fc00-405c-8c5d-d70e7cb132ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106725 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j9hz\" (UniqueName: \"kubernetes.io/projected/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-kube-api-access-4j9hz\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106785 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knvsv\" (UniqueName: \"kubernetes.io/projected/fc1bec83-d97b-403e-9de5-9f06087157ad-kube-api-access-knvsv\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106808 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-trusted-ca-bundle\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106831 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106884 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/001408b5-e454-432d-93dc-3a2dc497941f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106904 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5787789-e288-4011-baeb-17da6439f196-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.106924 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.107744 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fc1bec83-d97b-403e-9de5-9f06087157ad-trusted-ca\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.107821 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-service-ca\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.107887 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdqlc\" (UniqueName: \"kubernetes.io/projected/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-kube-api-access-tdqlc\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.107924 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.107957 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc1bec83-d97b-403e-9de5-9f06087157ad-serving-cert\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108003 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-config\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108085 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xm49\" (UniqueName: \"kubernetes.io/projected/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-kube-api-access-9xm49\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108151 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-metrics-certs\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108181 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-config\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108253 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82t8v\" (UniqueName: \"kubernetes.io/projected/3e1f11da-af03-4bc8-97c8-ad2eac21d489-kube-api-access-82t8v\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108904 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e5bf21-b590-45a3-9891-3e5114d6b25d-serving-cert\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.108977 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-serving-cert\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109015 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-bound-sa-token\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109039 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109063 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-ca\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109112 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z42jj\" (UniqueName: \"kubernetes.io/projected/b3206d6a-5f6e-4f19-9dac-93b2689734dc-kube-api-access-z42jj\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109134 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1bec83-d97b-403e-9de5-9f06087157ad-config\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109157 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-dir\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109205 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5787789-e288-4011-baeb-17da6439f196-config\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109224 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ba6778c-5684-46a3-b7c5-75f26973e3e0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109252 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-registry-tls\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109274 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109306 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109332 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109349 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109372 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-oauth-serving-cert\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109394 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9021090f-c067-4574-979a-4186a2df4b89-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-96m8f\" (UID: \"9021090f-c067-4574-979a-4186a2df4b89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109443 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-registry-certificates\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109463 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.109488 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.110104 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:10.610081697 +0000 UTC m=+147.664980539 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.115554 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.167927 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.174836 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211120 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211614 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r66l\" (UniqueName: \"kubernetes.io/projected/f7c84c47-7382-42cd-92b0-5e0261774961-kube-api-access-5r66l\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211665 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q22dx\" (UniqueName: \"kubernetes.io/projected/7b29632b-7b0b-47ee-b411-87c48881db23-kube-api-access-q22dx\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211701 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-csi-data-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211732 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3c85703d-4a60-4076-b96f-f55fe7483527-node-bootstrap-token\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.211766 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:10.711735615 +0000 UTC m=+147.766634457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211859 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-service-ca\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211885 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211914 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2clm\" (UniqueName: \"kubernetes.io/projected/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-kube-api-access-p2clm\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211941 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f9a02538-c9f6-41ed-b022-89bafc7d548a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211960 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9brz\" (UniqueName: \"kubernetes.io/projected/acfd5bdb-80ee-455d-9c3a-765ab62a7689-kube-api-access-t9brz\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211980 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7b91e57c-19f9-49a5-9d92-6be098455d2b-images\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.211998 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498cb9d-681f-4351-b0a2-1606bcc3afdf-config\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212033 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-trusted-ca\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212050 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9498cb9d-681f-4351-b0a2-1606bcc3afdf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212066 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-mountpoint-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212084 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23705022-817e-4dc7-9303-b2ee99713d42-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212102 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5787789-e288-4011-baeb-17da6439f196-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212120 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212136 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-signing-cabundle\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212155 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fc1bec83-d97b-403e-9de5-9f06087157ad-trusted-ca\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212175 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-config\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212192 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b86de852-56af-4d31-b2cf-3c27bd5ff599-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212236 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-config\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212255 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv5fd\" (UniqueName: \"kubernetes.io/projected/cdf6f002-40c1-443b-a1fe-89f312da84f2-kube-api-access-pv5fd\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212275 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-metrics-tls\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212316 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e5bf21-b590-45a3-9891-3e5114d6b25d-serving-cert\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212336 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-bound-sa-token\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212355 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212380 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-ca\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212397 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdf6f002-40c1-443b-a1fe-89f312da84f2-config-volume\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212415 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1bec83-d97b-403e-9de5-9f06087157ad-config\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212434 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-dir\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212450 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbvrp\" (UniqueName: \"kubernetes.io/projected/5ff75676-97f2-4f72-bfbe-4dfd756310d9-kube-api-access-zbvrp\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212470 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5787789-e288-4011-baeb-17da6439f196-config\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212487 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ba6778c-5684-46a3-b7c5-75f26973e3e0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212506 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212523 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9a02538-c9f6-41ed-b022-89bafc7d548a-srv-cert\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212541 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86de852-56af-4d31-b2cf-3c27bd5ff599-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212562 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-oauth-serving-cert\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212580 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-registry-certificates\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212596 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212615 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-config\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212647 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h44qb\" (UniqueName: \"kubernetes.io/projected/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-kube-api-access-h44qb\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212665 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3206d6a-5f6e-4f19-9dac-93b2689734dc-serving-cert\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212681 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3c85703d-4a60-4076-b96f-f55fe7483527-certs\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212696 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-plugins-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212728 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ba6778c-5684-46a3-b7c5-75f26973e3e0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212745 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ff75676-97f2-4f72-bfbe-4dfd756310d9-webhook-cert\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.212762 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213029 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-signing-key\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213046 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/18f2bc53-5f87-46d7-ab1e-2d878b364a46-cert\") pod \"ingress-canary-4786l\" (UID: \"18f2bc53-5f87-46d7-ab1e-2d878b364a46\") " pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213062 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c84c47-7382-42cd-92b0-5e0261774961-serving-cert\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213084 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213105 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-default-certificate\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213125 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213147 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4vwr\" (UniqueName: \"kubernetes.io/projected/dc8c0c59-a185-4b91-920e-c53f58d27e61-kube-api-access-l4vwr\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213168 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thxfh\" (UniqueName: \"kubernetes.io/projected/b86de852-56af-4d31-b2cf-3c27bd5ff599-kube-api-access-thxfh\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213188 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7c2j\" (UniqueName: \"kubernetes.io/projected/35efc8ac-5b7d-4fbc-828d-cf84bb7d8649-kube-api-access-f7c2j\") pod \"multus-admission-controller-857f4d67dd-qv6pn\" (UID: \"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213216 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213237 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7b29632b-7b0b-47ee-b411-87c48881db23-srv-cert\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213258 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm5rb\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-kube-api-access-dm5rb\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213278 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8eefc13-080a-4cf2-928f-0289b856fba1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5gz2f\" (UID: \"d8eefc13-080a-4cf2-928f-0289b856fba1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213314 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbsz5\" (UniqueName: \"kubernetes.io/projected/e3e5bf21-b590-45a3-9891-3e5114d6b25d-kube-api-access-xbsz5\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213331 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7b29632b-7b0b-47ee-b411-87c48881db23-profile-collector-cert\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213371 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cdf6f002-40c1-443b-a1fe-89f312da84f2-secret-volume\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213393 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-registration-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213411 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddmnb\" (UniqueName: \"kubernetes.io/projected/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-kube-api-access-ddmnb\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213432 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-client-ca\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213448 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-socket-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213478 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-oauth-config\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213498 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpd47\" (UniqueName: \"kubernetes.io/projected/e2745959-fc00-405c-8c5d-d70e7cb132ec-kube-api-access-zpd47\") pod \"dns-operator-744455d44c-bbx6q\" (UID: \"e2745959-fc00-405c-8c5d-d70e7cb132ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213515 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-serving-cert\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213538 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5420a718-1d0a-4b72-b57d-a8a4a18e7763-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdbxg\" (UID: \"5420a718-1d0a-4b72-b57d-a8a4a18e7763\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213558 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95km6\" (UniqueName: \"kubernetes.io/projected/3c85703d-4a60-4076-b96f-f55fe7483527-kube-api-access-95km6\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213578 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j9hz\" (UniqueName: \"kubernetes.io/projected/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-kube-api-access-4j9hz\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213598 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2745959-fc00-405c-8c5d-d70e7cb132ec-metrics-tls\") pod \"dns-operator-744455d44c-bbx6q\" (UID: \"e2745959-fc00-405c-8c5d-d70e7cb132ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213615 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9498cb9d-681f-4351-b0a2-1606bcc3afdf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213699 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9hsw\" (UniqueName: \"kubernetes.io/projected/18f2bc53-5f87-46d7-ab1e-2d878b364a46-kube-api-access-z9hsw\") pod \"ingress-canary-4786l\" (UID: \"18f2bc53-5f87-46d7-ab1e-2d878b364a46\") " pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213719 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knvsv\" (UniqueName: \"kubernetes.io/projected/fc1bec83-d97b-403e-9de5-9f06087157ad-kube-api-access-knvsv\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213736 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-trusted-ca-bundle\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213753 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213772 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/001408b5-e454-432d-93dc-3a2dc497941f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213791 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n49vm\" (UniqueName: \"kubernetes.io/projected/d8eefc13-080a-4cf2-928f-0289b856fba1-kube-api-access-n49vm\") pod \"package-server-manager-789f6589d5-5gz2f\" (UID: \"d8eefc13-080a-4cf2-928f-0289b856fba1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213810 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-service-ca\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213830 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdqlc\" (UniqueName: \"kubernetes.io/projected/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-kube-api-access-tdqlc\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213850 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213866 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/35efc8ac-5b7d-4fbc-828d-cf84bb7d8649-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qv6pn\" (UID: \"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213882 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc1bec83-d97b-403e-9de5-9f06087157ad-serving-cert\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213926 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xm49\" (UniqueName: \"kubernetes.io/projected/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-kube-api-access-9xm49\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213942 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acfd5bdb-80ee-455d-9c3a-765ab62a7689-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213958 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-metrics-certs\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213981 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82t8v\" (UniqueName: \"kubernetes.io/projected/3e1f11da-af03-4bc8-97c8-ad2eac21d489-kube-api-access-82t8v\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.213998 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-serving-cert\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214015 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b91e57c-19f9-49a5-9d92-6be098455d2b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214033 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z42jj\" (UniqueName: \"kubernetes.io/projected/b3206d6a-5f6e-4f19-9dac-93b2689734dc-kube-api-access-z42jj\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214048 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs64l\" (UniqueName: \"kubernetes.io/projected/f9a02538-c9f6-41ed-b022-89bafc7d548a-kube-api-access-vs64l\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214065 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b91e57c-19f9-49a5-9d92-6be098455d2b-proxy-tls\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214098 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-registry-tls\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214113 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23705022-817e-4dc7-9303-b2ee99713d42-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214133 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214151 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214167 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214184 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgbz9\" (UniqueName: \"kubernetes.io/projected/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-kube-api-access-mgbz9\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214203 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9021090f-c067-4574-979a-4186a2df4b89-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-96m8f\" (UID: \"9021090f-c067-4574-979a-4186a2df4b89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214218 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214248 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c84c47-7382-42cd-92b0-5e0261774961-config\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214268 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214291 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214314 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214337 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ff75676-97f2-4f72-bfbe-4dfd756310d9-apiservice-cert\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214356 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2b4v\" (UniqueName: \"kubernetes.io/projected/9021090f-c067-4574-979a-4186a2df4b89-kube-api-access-p2b4v\") pod \"cluster-samples-operator-665b6dd947-96m8f\" (UID: \"9021090f-c067-4574-979a-4186a2df4b89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214373 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-policies\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214393 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/001408b5-e454-432d-93dc-3a2dc497941f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214410 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmx8b\" (UniqueName: \"kubernetes.io/projected/d5787789-e288-4011-baeb-17da6439f196-kube-api-access-jmx8b\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214438 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-client\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214454 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdrcb\" (UniqueName: \"kubernetes.io/projected/7b91e57c-19f9-49a5-9d92-6be098455d2b-kube-api-access-kdrcb\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214472 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5787789-e288-4011-baeb-17da6439f196-images\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214489 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ba6778c-5684-46a3-b7c5-75f26973e3e0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214506 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-config-volume\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214522 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpqrq\" (UniqueName: \"kubernetes.io/projected/d8c9427b-899a-49be-8253-369d7c853822-kube-api-access-tpqrq\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214538 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-stats-auth\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214555 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggmhj\" (UniqueName: \"kubernetes.io/projected/5420a718-1d0a-4b72-b57d-a8a4a18e7763-kube-api-access-ggmhj\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdbxg\" (UID: \"5420a718-1d0a-4b72-b57d-a8a4a18e7763\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214570 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5ff75676-97f2-4f72-bfbe-4dfd756310d9-tmpfs\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214585 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.215779 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fc1bec83-d97b-403e-9de5-9f06087157ad-trusted-ca\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.214615 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.219271 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23705022-817e-4dc7-9303-b2ee99713d42-config\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.219302 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/acfd5bdb-80ee-455d-9c3a-765ab62a7689-proxy-tls\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.219339 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e1f11da-af03-4bc8-97c8-ad2eac21d489-service-ca-bundle\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.219853 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-config\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.220063 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5787789-e288-4011-baeb-17da6439f196-config\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.220414 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-service-ca\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.220542 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.221687 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-trusted-ca\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.222131 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-ca\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.222154 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-dir\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.222439 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-trusted-ca-bundle\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.222646 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/001408b5-e454-432d-93dc-3a2dc497941f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.223617 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.224060 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.224789 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1bec83-d97b-403e-9de5-9f06087157ad-config\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.225330 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-config\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.225391 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e1f11da-af03-4bc8-97c8-ad2eac21d489-service-ca-bundle\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.225960 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.226226 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-config\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.226868 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dc8c0c59-a185-4b91-920e-c53f58d27e61-oauth-serving-cert\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.227789 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-registry-certificates\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.228858 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.230063 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.230752 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e2745959-fc00-405c-8c5d-d70e7cb132ec-metrics-tls\") pod \"dns-operator-744455d44c-bbx6q\" (UID: \"e2745959-fc00-405c-8c5d-d70e7cb132ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.232086 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-policies\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.232576 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e5bf21-b590-45a3-9891-3e5114d6b25d-serving-cert\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.233120 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d5787789-e288-4011-baeb-17da6439f196-images\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.233562 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9021090f-c067-4574-979a-4186a2df4b89-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-96m8f\" (UID: \"9021090f-c067-4574-979a-4186a2df4b89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.233679 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-client-ca\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.233946 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:10.733926804 +0000 UTC m=+147.788825646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.234055 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3206d6a-5f6e-4f19-9dac-93b2689734dc-serving-cert\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.234576 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ba6778c-5684-46a3-b7c5-75f26973e3e0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.234641 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-service-ca\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.235093 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-registry-tls\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.235809 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.236030 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.236392 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-oauth-config\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.237877 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/001408b5-e454-432d-93dc-3a2dc497941f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.238291 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.238811 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-default-certificate\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.239582 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc1bec83-d97b-403e-9de5-9f06087157ad-serving-cert\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.239828 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ba6778c-5684-46a3-b7c5-75f26973e3e0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.243809 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.244109 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.244597 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d5787789-e288-4011-baeb-17da6439f196-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.245057 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-stats-auth\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.248353 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5420a718-1d0a-4b72-b57d-a8a4a18e7763-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdbxg\" (UID: \"5420a718-1d0a-4b72-b57d-a8a4a18e7763\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.248761 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.249455 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.249798 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3e1f11da-af03-4bc8-97c8-ad2eac21d489-metrics-certs\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.251247 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e3e5bf21-b590-45a3-9891-3e5114d6b25d-etcd-client\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.251269 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.251878 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.252812 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-serving-cert\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.253436 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc8c0c59-a185-4b91-920e-c53f58d27e61-console-serving-cert\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.254842 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.257076 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-bound-sa-token\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.281147 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.298906 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4vwr\" (UniqueName: \"kubernetes.io/projected/dc8c0c59-a185-4b91-920e-c53f58d27e61-kube-api-access-l4vwr\") pod \"console-f9d7485db-wfp5w\" (UID: \"dc8c0c59-a185-4b91-920e-c53f58d27e61\") " pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.313908 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.320828 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321157 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3c85703d-4a60-4076-b96f-f55fe7483527-certs\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321190 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-plugins-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321214 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ff75676-97f2-4f72-bfbe-4dfd756310d9-webhook-cert\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321234 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-signing-key\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321256 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/18f2bc53-5f87-46d7-ab1e-2d878b364a46-cert\") pod \"ingress-canary-4786l\" (UID: \"18f2bc53-5f87-46d7-ab1e-2d878b364a46\") " pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321273 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c84c47-7382-42cd-92b0-5e0261774961-serving-cert\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321299 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thxfh\" (UniqueName: \"kubernetes.io/projected/b86de852-56af-4d31-b2cf-3c27bd5ff599-kube-api-access-thxfh\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321321 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7c2j\" (UniqueName: \"kubernetes.io/projected/35efc8ac-5b7d-4fbc-828d-cf84bb7d8649-kube-api-access-f7c2j\") pod \"multus-admission-controller-857f4d67dd-qv6pn\" (UID: \"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321346 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7b29632b-7b0b-47ee-b411-87c48881db23-srv-cert\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321373 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8eefc13-080a-4cf2-928f-0289b856fba1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5gz2f\" (UID: \"d8eefc13-080a-4cf2-928f-0289b856fba1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321405 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7b29632b-7b0b-47ee-b411-87c48881db23-profile-collector-cert\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321425 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cdf6f002-40c1-443b-a1fe-89f312da84f2-secret-volume\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321449 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-socket-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321466 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-registration-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321491 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95km6\" (UniqueName: \"kubernetes.io/projected/3c85703d-4a60-4076-b96f-f55fe7483527-kube-api-access-95km6\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321537 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9498cb9d-681f-4351-b0a2-1606bcc3afdf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321558 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9hsw\" (UniqueName: \"kubernetes.io/projected/18f2bc53-5f87-46d7-ab1e-2d878b364a46-kube-api-access-z9hsw\") pod \"ingress-canary-4786l\" (UID: \"18f2bc53-5f87-46d7-ab1e-2d878b364a46\") " pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321579 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n49vm\" (UniqueName: \"kubernetes.io/projected/d8eefc13-080a-4cf2-928f-0289b856fba1-kube-api-access-n49vm\") pod \"package-server-manager-789f6589d5-5gz2f\" (UID: \"d8eefc13-080a-4cf2-928f-0289b856fba1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321635 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/35efc8ac-5b7d-4fbc-828d-cf84bb7d8649-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qv6pn\" (UID: \"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321667 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acfd5bdb-80ee-455d-9c3a-765ab62a7689-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321699 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b91e57c-19f9-49a5-9d92-6be098455d2b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321723 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs64l\" (UniqueName: \"kubernetes.io/projected/f9a02538-c9f6-41ed-b022-89bafc7d548a-kube-api-access-vs64l\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321741 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b91e57c-19f9-49a5-9d92-6be098455d2b-proxy-tls\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321767 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgbz9\" (UniqueName: \"kubernetes.io/projected/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-kube-api-access-mgbz9\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321783 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23705022-817e-4dc7-9303-b2ee99713d42-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321802 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c84c47-7382-42cd-92b0-5e0261774961-config\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321822 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.321839 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ff75676-97f2-4f72-bfbe-4dfd756310d9-apiservice-cert\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322643 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdrcb\" (UniqueName: \"kubernetes.io/projected/7b91e57c-19f9-49a5-9d92-6be098455d2b-kube-api-access-kdrcb\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322684 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-config-volume\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322706 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpqrq\" (UniqueName: \"kubernetes.io/projected/d8c9427b-899a-49be-8253-369d7c853822-kube-api-access-tpqrq\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322741 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5ff75676-97f2-4f72-bfbe-4dfd756310d9-tmpfs\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322765 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322781 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23705022-817e-4dc7-9303-b2ee99713d42-config\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322799 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/acfd5bdb-80ee-455d-9c3a-765ab62a7689-proxy-tls\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322818 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q22dx\" (UniqueName: \"kubernetes.io/projected/7b29632b-7b0b-47ee-b411-87c48881db23-kube-api-access-q22dx\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322835 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-csi-data-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322854 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r66l\" (UniqueName: \"kubernetes.io/projected/f7c84c47-7382-42cd-92b0-5e0261774961-kube-api-access-5r66l\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322871 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3c85703d-4a60-4076-b96f-f55fe7483527-node-bootstrap-token\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322895 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2clm\" (UniqueName: \"kubernetes.io/projected/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-kube-api-access-p2clm\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322912 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f9a02538-c9f6-41ed-b022-89bafc7d548a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322930 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9brz\" (UniqueName: \"kubernetes.io/projected/acfd5bdb-80ee-455d-9c3a-765ab62a7689-kube-api-access-t9brz\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322950 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7b91e57c-19f9-49a5-9d92-6be098455d2b-images\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322967 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498cb9d-681f-4351-b0a2-1606bcc3afdf-config\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.322990 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9498cb9d-681f-4351-b0a2-1606bcc3afdf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323008 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-mountpoint-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323025 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-signing-cabundle\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323047 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23705022-817e-4dc7-9303-b2ee99713d42-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323077 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b86de852-56af-4d31-b2cf-3c27bd5ff599-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323096 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv5fd\" (UniqueName: \"kubernetes.io/projected/cdf6f002-40c1-443b-a1fe-89f312da84f2-kube-api-access-pv5fd\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323113 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-metrics-tls\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323150 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdf6f002-40c1-443b-a1fe-89f312da84f2-config-volume\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323171 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbvrp\" (UniqueName: \"kubernetes.io/projected/5ff75676-97f2-4f72-bfbe-4dfd756310d9-kube-api-access-zbvrp\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323187 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9a02538-c9f6-41ed-b022-89bafc7d548a-srv-cert\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323206 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86de852-56af-4d31-b2cf-3c27bd5ff599-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.323224 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h44qb\" (UniqueName: \"kubernetes.io/projected/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-kube-api-access-h44qb\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.323535 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:10.823514754 +0000 UTC m=+147.878413596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.324429 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-registration-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.324522 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-plugins-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.325034 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-socket-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.325904 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b91e57c-19f9-49a5-9d92-6be098455d2b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.327275 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knvsv\" (UniqueName: \"kubernetes.io/projected/fc1bec83-d97b-403e-9de5-9f06087157ad-kube-api-access-knvsv\") pod \"console-operator-58897d9998-rgcdb\" (UID: \"fc1bec83-d97b-403e-9de5-9f06087157ad\") " pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.333475 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w4dpg"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.333736 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7b29632b-7b0b-47ee-b411-87c48881db23-srv-cert\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.333854 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cdf6f002-40c1-443b-a1fe-89f312da84f2-secret-volume\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.333950 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-signing-key\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.334592 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/acfd5bdb-80ee-455d-9c3a-765ab62a7689-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.334748 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-config-volume\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.334815 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/35efc8ac-5b7d-4fbc-828d-cf84bb7d8649-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qv6pn\" (UID: \"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.335018 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7c84c47-7382-42cd-92b0-5e0261774961-config\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.335392 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5ff75676-97f2-4f72-bfbe-4dfd756310d9-tmpfs\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.335612 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5ff75676-97f2-4f72-bfbe-4dfd756310d9-apiservice-cert\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.336045 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-mountpoint-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.336066 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9498cb9d-681f-4351-b0a2-1606bcc3afdf-config\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.336380 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdf6f002-40c1-443b-a1fe-89f312da84f2-config-volume\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.336386 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3c85703d-4a60-4076-b96f-f55fe7483527-certs\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.336735 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5ff75676-97f2-4f72-bfbe-4dfd756310d9-webhook-cert\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.336995 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86de852-56af-4d31-b2cf-3c27bd5ff599-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.337672 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/18f2bc53-5f87-46d7-ab1e-2d878b364a46-cert\") pod \"ingress-canary-4786l\" (UID: \"18f2bc53-5f87-46d7-ab1e-2d878b364a46\") " pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.337719 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-signing-cabundle\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.337799 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23705022-817e-4dc7-9303-b2ee99713d42-config\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.337900 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d8c9427b-899a-49be-8253-369d7c853822-csi-data-dir\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.338558 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7c84c47-7382-42cd-92b0-5e0261774961-serving-cert\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.338896 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.339206 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23705022-817e-4dc7-9303-b2ee99713d42-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.339282 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdqlc\" (UniqueName: \"kubernetes.io/projected/c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97-kube-api-access-tdqlc\") pod \"cluster-image-registry-operator-dc59b4c8b-xqmwx\" (UID: \"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.339369 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7b29632b-7b0b-47ee-b411-87c48881db23-profile-collector-cert\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.340315 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b86de852-56af-4d31-b2cf-3c27bd5ff599-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.341180 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d8eefc13-080a-4cf2-928f-0289b856fba1-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5gz2f\" (UID: \"d8eefc13-080a-4cf2-928f-0289b856fba1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.341337 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-metrics-tls\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.341407 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.341409 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b91e57c-19f9-49a5-9d92-6be098455d2b-proxy-tls\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.342082 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.344486 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/acfd5bdb-80ee-455d-9c3a-765ab62a7689-proxy-tls\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.346306 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9498cb9d-681f-4351-b0a2-1606bcc3afdf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: W1004 07:12:10.360147 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode435b9eb_8082_4935_8d71_14ea61d5119a.slice/crio-5ee373cba514cac84bd49b98022061aa62590dcef69f4d09702beb87b2914765 WatchSource:0}: Error finding container 5ee373cba514cac84bd49b98022061aa62590dcef69f4d09702beb87b2914765: Status 404 returned error can't find the container with id 5ee373cba514cac84bd49b98022061aa62590dcef69f4d09702beb87b2914765 Oct 04 07:12:10 crc kubenswrapper[4987]: W1004 07:12:10.362304 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e6f14f4_f9b0_438a_a4b4_290b3beb7814.slice/crio-6144451132b645a53c1952af8f3ca60cf8d1e784976c7b1f00a9caa0eadc5f2f WatchSource:0}: Error finding container 6144451132b645a53c1952af8f3ca60cf8d1e784976c7b1f00a9caa0eadc5f2f: Status 404 returned error can't find the container with id 6144451132b645a53c1952af8f3ca60cf8d1e784976c7b1f00a9caa0eadc5f2f Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.363299 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm5rb\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-kube-api-access-dm5rb\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.374206 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7b91e57c-19f9-49a5-9d92-6be098455d2b-images\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.374833 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f9a02538-c9f6-41ed-b022-89bafc7d548a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.374893 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9a02538-c9f6-41ed-b022-89bafc7d548a-srv-cert\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.379150 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbsz5\" (UniqueName: \"kubernetes.io/projected/e3e5bf21-b590-45a3-9891-3e5114d6b25d-kube-api-access-xbsz5\") pod \"etcd-operator-b45778765-5b448\" (UID: \"e3e5bf21-b590-45a3-9891-3e5114d6b25d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.383503 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3c85703d-4a60-4076-b96f-f55fe7483527-node-bootstrap-token\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.391645 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2b4v\" (UniqueName: \"kubernetes.io/projected/9021090f-c067-4574-979a-4186a2df4b89-kube-api-access-p2b4v\") pod \"cluster-samples-operator-665b6dd947-96m8f\" (UID: \"9021090f-c067-4574-979a-4186a2df4b89\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.407527 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.412306 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z42jj\" (UniqueName: \"kubernetes.io/projected/b3206d6a-5f6e-4f19-9dac-93b2689734dc-kube-api-access-z42jj\") pod \"route-controller-manager-6576b87f9c-cp8x9\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.426920 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.427451 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:10.927436908 +0000 UTC m=+147.982335750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.434967 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmx8b\" (UniqueName: \"kubernetes.io/projected/d5787789-e288-4011-baeb-17da6439f196-kube-api-access-jmx8b\") pod \"machine-api-operator-5694c8668f-f49kw\" (UID: \"d5787789-e288-4011-baeb-17da6439f196\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.438757 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.446846 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.458444 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-wt287"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.464823 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpd47\" (UniqueName: \"kubernetes.io/projected/e2745959-fc00-405c-8c5d-d70e7cb132ec-kube-api-access-zpd47\") pod \"dns-operator-744455d44c-bbx6q\" (UID: \"e2745959-fc00-405c-8c5d-d70e7cb132ec\") " pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.471577 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.475505 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tfd59"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.477528 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddmnb\" (UniqueName: \"kubernetes.io/projected/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-kube-api-access-ddmnb\") pod \"oauth-openshift-558db77b4-6hddn\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.482986 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.494989 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggmhj\" (UniqueName: \"kubernetes.io/projected/5420a718-1d0a-4b72-b57d-a8a4a18e7763-kube-api-access-ggmhj\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdbxg\" (UID: \"5420a718-1d0a-4b72-b57d-a8a4a18e7763\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.509216 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.517939 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xm49\" (UniqueName: \"kubernetes.io/projected/b475bd2a-4652-40aa-8e9f-0f4a069ad77f-kube-api-access-9xm49\") pod \"openshift-config-operator-7777fb866f-f4gb9\" (UID: \"b475bd2a-4652-40aa-8e9f-0f4a069ad77f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.529297 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.535374 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.035322056 +0000 UTC m=+148.090220898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.536241 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.536913 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.036890397 +0000 UTC m=+148.091789239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.549496 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wfp5w"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.550274 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82t8v\" (UniqueName: \"kubernetes.io/projected/3e1f11da-af03-4bc8-97c8-ad2eac21d489-kube-api-access-82t8v\") pod \"router-default-5444994796-bjgbf\" (UID: \"3e1f11da-af03-4bc8-97c8-ad2eac21d489\") " pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.571992 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ba6778c-5684-46a3-b7c5-75f26973e3e0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jsh2d\" (UID: \"7ba6778c-5684-46a3-b7c5-75f26973e3e0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.591258 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j9hz\" (UniqueName: \"kubernetes.io/projected/fc3745bd-0124-4a55-b9b3-a7b27b1aee2a-kube-api-access-4j9hz\") pod \"openshift-apiserver-operator-796bbdcf4f-2ggpf\" (UID: \"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.597837 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.618513 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h44qb\" (UniqueName: \"kubernetes.io/projected/bbe932ef-c8f2-4eba-a23b-226aaf51ceaa-kube-api-access-h44qb\") pod \"dns-default-nv9l8\" (UID: \"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa\") " pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.625043 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.626101 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.626664 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-t2w8v"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.634552 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7c2j\" (UniqueName: \"kubernetes.io/projected/35efc8ac-5b7d-4fbc-828d-cf84bb7d8649-kube-api-access-f7c2j\") pod \"multus-admission-controller-857f4d67dd-qv6pn\" (UID: \"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.637369 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.637712 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.137678219 +0000 UTC m=+148.192577071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.638062 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.638595 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.138577055 +0000 UTC m=+148.193475897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.655394 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95km6\" (UniqueName: \"kubernetes.io/projected/3c85703d-4a60-4076-b96f-f55fe7483527-kube-api-access-95km6\") pod \"machine-config-server-vzg8c\" (UID: \"3c85703d-4a60-4076-b96f-f55fe7483527\") " pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.659809 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.667945 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.673001 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vzg8c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.677447 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.677757 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9498cb9d-681f-4351-b0a2-1606bcc3afdf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7z9pt\" (UID: \"9498cb9d-681f-4351-b0a2-1606bcc3afdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.696590 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9hsw\" (UniqueName: \"kubernetes.io/projected/18f2bc53-5f87-46d7-ab1e-2d878b364a46-kube-api-access-z9hsw\") pod \"ingress-canary-4786l\" (UID: \"18f2bc53-5f87-46d7-ab1e-2d878b364a46\") " pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.703475 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.713325 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n49vm\" (UniqueName: \"kubernetes.io/projected/d8eefc13-080a-4cf2-928f-0289b856fba1-kube-api-access-n49vm\") pod \"package-server-manager-789f6589d5-5gz2f\" (UID: \"d8eefc13-080a-4cf2-928f-0289b856fba1\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.735310 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs64l\" (UniqueName: \"kubernetes.io/projected/f9a02538-c9f6-41ed-b022-89bafc7d548a-kube-api-access-vs64l\") pod \"olm-operator-6b444d44fb-nv4lt\" (UID: \"f9a02538-c9f6-41ed-b022-89bafc7d548a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.739355 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.739838 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.239821225 +0000 UTC m=+148.294720067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.744410 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" event={"ID":"3b67efca-532b-437c-b377-2c3bfb7372c0","Type":"ContainerStarted","Data":"3d45d435dc8524b1263f7a93d70fd96f393fc7654315967ef3c57975951d9ced"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.745287 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.746054 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" event={"ID":"fb9f070f-4908-41d2-b247-f0805b1c4792","Type":"ContainerStarted","Data":"31e35521a645c0673ddb2609279e675117ae6dc6333b6a4e92e56cdd2ff8478a"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.746834 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt287" event={"ID":"262ca319-a044-47bb-b107-1b76e9d9abbb","Type":"ContainerStarted","Data":"6419f708d04374d1ab9647aa29ea18948f410b8da73b04aa8e4a401a7677f4a3"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.747516 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" event={"ID":"1a5130dd-df25-4642-8180-637b14f8ee4a","Type":"ContainerStarted","Data":"5197262856d44f062f99a35252a3c26217bd724053d9ff5f55ffceea0c36e70e"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.755340 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.755754 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" event={"ID":"e435b9eb-8082-4935-8d71-14ea61d5119a","Type":"ContainerStarted","Data":"5ee373cba514cac84bd49b98022061aa62590dcef69f4d09702beb87b2914765"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.762136 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thxfh\" (UniqueName: \"kubernetes.io/projected/b86de852-56af-4d31-b2cf-3c27bd5ff599-kube-api-access-thxfh\") pod \"kube-storage-version-migrator-operator-b67b599dd-c2hns\" (UID: \"b86de852-56af-4d31-b2cf-3c27bd5ff599\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.771873 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wfp5w" event={"ID":"dc8c0c59-a185-4b91-920e-c53f58d27e61","Type":"ContainerStarted","Data":"b1fdcf1e2f59bd4bc8d07e33790bc0a040de28d0d7bff37b258a86d35ad6cc25"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.773368 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-t2w8v" event={"ID":"453b9a94-8a64-477b-97eb-633594674f79","Type":"ContainerStarted","Data":"ba0b67b2e08d269ceb5ce3a5980766b36ca0f114d3fe11b731374b9c75f130ef"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.774641 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" event={"ID":"25e66a14-b32b-4aca-9fce-9fc81f5d66f4","Type":"ContainerStarted","Data":"9f937282b0f29c8c9e967b0a21b2b8a06292bec3f50a0dd96da7b3bb33748891"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.774667 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" event={"ID":"25e66a14-b32b-4aca-9fce-9fc81f5d66f4","Type":"ContainerStarted","Data":"ba5ad637ec71cc2f4f56d1c9b970a816aaf950c5827ce9f03fd2f316f6573f56"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.776098 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpqrq\" (UniqueName: \"kubernetes.io/projected/d8c9427b-899a-49be-8253-369d7c853822-kube-api-access-tpqrq\") pod \"csi-hostpathplugin-pdb9r\" (UID: \"d8c9427b-899a-49be-8253-369d7c853822\") " pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.779077 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" event={"ID":"acc7a457-28b0-4bd8-98aa-85110a95ae02","Type":"ContainerStarted","Data":"c6951d2f6496ae5d3df2367c3ecb13d66f5cc6e09bece0af01805882d19773fa"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.794740 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2clm\" (UniqueName: \"kubernetes.io/projected/ffcf2982-cee1-49a6-a2c4-15b23b45ad8b-kube-api-access-p2clm\") pod \"service-ca-9c57cc56f-zzpsj\" (UID: \"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b\") " pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.801800 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" event={"ID":"8e6f14f4-f9b0-438a-a4b4-290b3beb7814","Type":"ContainerStarted","Data":"6144451132b645a53c1952af8f3ca60cf8d1e784976c7b1f00a9caa0eadc5f2f"} Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.816963 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.817600 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q22dx\" (UniqueName: \"kubernetes.io/projected/7b29632b-7b0b-47ee-b411-87c48881db23-kube-api-access-q22dx\") pod \"catalog-operator-68c6474976-7sbxk\" (UID: \"7b29632b-7b0b-47ee-b411-87c48881db23\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.827034 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.841120 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5b448"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.841437 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.841758 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.341746223 +0000 UTC m=+148.396645065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.845731 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdrcb\" (UniqueName: \"kubernetes.io/projected/7b91e57c-19f9-49a5-9d92-6be098455d2b-kube-api-access-kdrcb\") pod \"machine-config-operator-74547568cd-v9cn9\" (UID: \"7b91e57c-19f9-49a5-9d92-6be098455d2b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.853187 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.856399 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbvrp\" (UniqueName: \"kubernetes.io/projected/5ff75676-97f2-4f72-bfbe-4dfd756310d9-kube-api-access-zbvrp\") pod \"packageserver-d55dfcdfc-vw86c\" (UID: \"5ff75676-97f2-4f72-bfbe-4dfd756310d9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.864184 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.869977 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgbz9\" (UniqueName: \"kubernetes.io/projected/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-kube-api-access-mgbz9\") pod \"marketplace-operator-79b997595-j9wm6\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.884369 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.895352 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.901339 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.905084 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv5fd\" (UniqueName: \"kubernetes.io/projected/cdf6f002-40c1-443b-a1fe-89f312da84f2-kube-api-access-pv5fd\") pod \"collect-profiles-29326020-h7n99\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.921005 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.923593 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.927472 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.933050 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.933762 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23705022-817e-4dc7-9303-b2ee99713d42-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-b5z4v\" (UID: \"23705022-817e-4dc7-9303-b2ee99713d42\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.934038 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.936920 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r66l\" (UniqueName: \"kubernetes.io/projected/f7c84c47-7382-42cd-92b0-5e0261774961-kube-api-access-5r66l\") pod \"service-ca-operator-777779d784-vtxt4\" (UID: \"f7c84c47-7382-42cd-92b0-5e0261774961\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.943477 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.943771 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:10 crc kubenswrapper[4987]: E1004 07:12:10.944286 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.444266483 +0000 UTC m=+148.499165325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.946234 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rgcdb"] Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.949232 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4786l" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.957957 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.963020 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9brz\" (UniqueName: \"kubernetes.io/projected/acfd5bdb-80ee-455d-9c3a-765ab62a7689-kube-api-access-t9brz\") pod \"machine-config-controller-84d6567774-c4hjg\" (UID: \"acfd5bdb-80ee-455d-9c3a-765ab62a7689\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.963231 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" Oct 04 07:12:10 crc kubenswrapper[4987]: I1004 07:12:10.993957 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.004022 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.045173 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.045575 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.545557746 +0000 UTC m=+148.600456588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.066038 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3e5bf21_b590_45a3_9891_3e5114d6b25d.slice/crio-3c6cb7bff2e03f8f055e45d5317a5e2ad36614a206b6c8513d80e619281b73b7 WatchSource:0}: Error finding container 3c6cb7bff2e03f8f055e45d5317a5e2ad36614a206b6c8513d80e619281b73b7: Status 404 returned error can't find the container with id 3c6cb7bff2e03f8f055e45d5317a5e2ad36614a206b6c8513d80e619281b73b7 Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.068522 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3206d6a_5f6e_4f19_9dac_93b2689734dc.slice/crio-f0a98fae1a2c3c121c8a125215f9aaf52e80201f856366ac4b81e665c1dd2e0f WatchSource:0}: Error finding container f0a98fae1a2c3c121c8a125215f9aaf52e80201f856366ac4b81e665c1dd2e0f: Status 404 returned error can't find the container with id f0a98fae1a2c3c121c8a125215f9aaf52e80201f856366ac4b81e665c1dd2e0f Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.073774 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc1bec83_d97b_403e_9de5_9f06087157ad.slice/crio-2430c2543f00f1f0b526af24f8f5746afd481675555c3edfaa581ffea3a04212 WatchSource:0}: Error finding container 2430c2543f00f1f0b526af24f8f5746afd481675555c3edfaa581ffea3a04212: Status 404 returned error can't find the container with id 2430c2543f00f1f0b526af24f8f5746afd481675555c3edfaa581ffea3a04212 Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.076096 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb475bd2a_4652_40aa_8e9f_0f4a069ad77f.slice/crio-9af95aae036a3de30d180a0698ebcb661fa99a89f4ea1ad0f1cd812c60a50a9e WatchSource:0}: Error finding container 9af95aae036a3de30d180a0698ebcb661fa99a89f4ea1ad0f1cd812c60a50a9e: Status 404 returned error can't find the container with id 9af95aae036a3de30d180a0698ebcb661fa99a89f4ea1ad0f1cd812c60a50a9e Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.076956 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3ff5bd7_52fa_4f8a_9dcb_2a8941ca1c97.slice/crio-d56ca500644d8059296da6199e51fc19947e0ee283d5847ca4fb680f57af8463 WatchSource:0}: Error finding container d56ca500644d8059296da6199e51fc19947e0ee283d5847ca4fb680f57af8463: Status 404 returned error can't find the container with id d56ca500644d8059296da6199e51fc19947e0ee283d5847ca4fb680f57af8463 Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.102131 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f49kw"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.137547 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bbx6q"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.143646 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.145905 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.146352 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.646334748 +0000 UTC m=+148.701233590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.173508 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.180226 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5787789_e288_4011_baeb_17da6439f196.slice/crio-01a61050069e242c3489c3dc39f462bbe3057a7aae6956421e74ff19673e176c WatchSource:0}: Error finding container 01a61050069e242c3489c3dc39f462bbe3057a7aae6956421e74ff19673e176c: Status 404 returned error can't find the container with id 01a61050069e242c3489c3dc39f462bbe3057a7aae6956421e74ff19673e176c Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.198849 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f"] Oct 04 07:12:11 crc kubenswrapper[4987]: W1004 07:12:11.208113 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c85703d_4a60_4076_b96f_f55fe7483527.slice/crio-951e2a4573129cb465fe4c37bd3c308e0becbb844666b5d8ef60187211bbbdfd WatchSource:0}: Error finding container 951e2a4573129cb465fe4c37bd3c308e0becbb844666b5d8ef60187211bbbdfd: Status 404 returned error can't find the container with id 951e2a4573129cb465fe4c37bd3c308e0becbb844666b5d8ef60187211bbbdfd Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.209034 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.247802 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.248179 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.748167962 +0000 UTC m=+148.803066804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.349239 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.349410 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.849385762 +0000 UTC m=+148.904284604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.349556 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.349872 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.84985943 +0000 UTC m=+148.904758272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.387703 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6hddn"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.450564 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.450745 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.950716206 +0000 UTC m=+149.005615058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.451086 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.451722 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:11.951703194 +0000 UTC m=+149.006602036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.552067 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.552592 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.052576311 +0000 UTC m=+149.107475153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.561890 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.654079 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.654435 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.154422815 +0000 UTC m=+149.209321657 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.708865 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zzpsj"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.728861 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nv9l8"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.730834 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.751824 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.755384 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.755697 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.255683247 +0000 UTC m=+149.310582089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.809123 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.828973 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" event={"ID":"d5787789-e288-4011-baeb-17da6439f196","Type":"ContainerStarted","Data":"01a61050069e242c3489c3dc39f462bbe3057a7aae6956421e74ff19673e176c"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.856370 4987 generic.go:334] "Generic (PLEG): container finished" podID="8e6f14f4-f9b0-438a-a4b4-290b3beb7814" containerID="4f7b95d5fee0fbef5606b9605cfc0a946962bec7b2471cf546c5abdc0eb67822" exitCode=0 Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.856468 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" event={"ID":"8e6f14f4-f9b0-438a-a4b4-290b3beb7814","Type":"ContainerDied","Data":"4f7b95d5fee0fbef5606b9605cfc0a946962bec7b2471cf546c5abdc0eb67822"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.857216 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.857610 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.357594713 +0000 UTC m=+149.412493565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.861545 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bjgbf" event={"ID":"3e1f11da-af03-4bc8-97c8-ad2eac21d489","Type":"ContainerStarted","Data":"aef0f7957dd0724a6fbe00b6b79d806f9eb8af77f52d7e666f962d42c9077025"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.869252 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" event={"ID":"fc1bec83-d97b-403e-9de5-9f06087157ad","Type":"ContainerStarted","Data":"2430c2543f00f1f0b526af24f8f5746afd481675555c3edfaa581ffea3a04212"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.870475 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" event={"ID":"659b7013-28e6-4ccb-b2d6-247c1eff7d4f","Type":"ContainerStarted","Data":"5bd6250477f0e9ac51a75b641052bc9bfd3553a9dd86bd3d893d500f00822ecf"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.872841 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" event={"ID":"acc7a457-28b0-4bd8-98aa-85110a95ae02","Type":"ContainerStarted","Data":"e5351e787923325a9bbff318b289ae1d57a3baf7ce2976caee2049689f5eb5cf"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.873774 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vzg8c" event={"ID":"3c85703d-4a60-4076-b96f-f55fe7483527","Type":"ContainerStarted","Data":"951e2a4573129cb465fe4c37bd3c308e0becbb844666b5d8ef60187211bbbdfd"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.876611 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" event={"ID":"e2745959-fc00-405c-8c5d-d70e7cb132ec","Type":"ContainerStarted","Data":"dca50fdce0d85c84f3d4e1e9270e432c45ddfd64efca01dfb98dd7a622a6308c"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.896130 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" event={"ID":"e435b9eb-8082-4935-8d71-14ea61d5119a","Type":"ContainerStarted","Data":"077cdf47bee9064a4202ad2b61ffaf7d5f859d0617d7cbadd979b72e5ff12d21"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.899865 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" event={"ID":"5420a718-1d0a-4b72-b57d-a8a4a18e7763","Type":"ContainerStarted","Data":"5da1710bc88d6c26c6ae9112487fc95a3cf2b3157c48159200ccc2d6ad92f97d"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.908318 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt287" event={"ID":"262ca319-a044-47bb-b107-1b76e9d9abbb","Type":"ContainerStarted","Data":"cfb46d9758db9ee0456887712129863ad224feda62d1dc3b054794ee9f4d8501"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.914298 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qv6pn"] Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.919054 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" event={"ID":"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97","Type":"ContainerStarted","Data":"d56ca500644d8059296da6199e51fc19947e0ee283d5847ca4fb680f57af8463"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.934284 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" event={"ID":"b3206d6a-5f6e-4f19-9dac-93b2689734dc","Type":"ContainerStarted","Data":"f0a98fae1a2c3c121c8a125215f9aaf52e80201f856366ac4b81e665c1dd2e0f"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.938282 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" event={"ID":"b475bd2a-4652-40aa-8e9f-0f4a069ad77f","Type":"ContainerStarted","Data":"9af95aae036a3de30d180a0698ebcb661fa99a89f4ea1ad0f1cd812c60a50a9e"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.948768 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-t2w8v" event={"ID":"453b9a94-8a64-477b-97eb-633594674f79","Type":"ContainerStarted","Data":"b336c5abc2c86bc8ead715263104a3fa3650d56771605fceb967c3b115b2b249"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.955841 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" event={"ID":"e3e5bf21-b590-45a3-9891-3e5114d6b25d","Type":"ContainerStarted","Data":"3c6cb7bff2e03f8f055e45d5317a5e2ad36614a206b6c8513d80e619281b73b7"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.957706 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:11 crc kubenswrapper[4987]: E1004 07:12:11.958082 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.458067224 +0000 UTC m=+149.512966066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.958549 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" event={"ID":"1a5130dd-df25-4642-8180-637b14f8ee4a","Type":"ContainerStarted","Data":"d7e183cdfc60afa6a7d71f1996684fbfcfd2abb560856e02700fd43e51897590"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.961582 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" event={"ID":"fb9f070f-4908-41d2-b247-f0805b1c4792","Type":"ContainerStarted","Data":"802ba9b6c8f1f4f9199a36d3e484c2c76b9df6ee891a2e19f2b31016e12823c5"} Oct 04 07:12:11 crc kubenswrapper[4987]: I1004 07:12:11.988234 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt"] Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.061055 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.561034982 +0000 UTC m=+149.615933824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.060444 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.168475 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.168933 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.668913229 +0000 UTC m=+149.723812081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.173249 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.173751 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.673734166 +0000 UTC m=+149.728633008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.192189 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.275682 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.276290 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.776258806 +0000 UTC m=+149.831157648 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.377759 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.378178 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.878164823 +0000 UTC m=+149.933063665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.390120 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b29632b_7b0b_47ee_b411_87c48881db23.slice/crio-c2c470bacd840c85073b8858b0577c9e1d4099adc1fde92f7b2b5ce5fa6c10db WatchSource:0}: Error finding container c2c470bacd840c85073b8858b0577c9e1d4099adc1fde92f7b2b5ce5fa6c10db: Status 404 returned error can't find the container with id c2c470bacd840c85073b8858b0577c9e1d4099adc1fde92f7b2b5ce5fa6c10db Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.407986 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-w4dpg" podStartSLOduration=128.407968688 podStartE2EDuration="2m8.407968688s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:12.405391768 +0000 UTC m=+149.460290610" watchObservedRunningTime="2025-10-04 07:12:12.407968688 +0000 UTC m=+149.462867530" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.453123 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.478766 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.478919 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.978904735 +0000 UTC m=+150.033803577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.479124 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.479508 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:12.979497047 +0000 UTC m=+150.034395889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.525829 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pdb9r"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.529721 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt"] Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.561942 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b91e57c_19f9_49a5_9d92_6be098455d2b.slice/crio-59ff6fa210ea9e2dcb08f8e5366e447177cb2900cb87d7f73da8b0bdabaab61b WatchSource:0}: Error finding container 59ff6fa210ea9e2dcb08f8e5366e447177cb2900cb87d7f73da8b0bdabaab61b: Status 404 returned error can't find the container with id 59ff6fa210ea9e2dcb08f8e5366e447177cb2900cb87d7f73da8b0bdabaab61b Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.574387 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8c9427b_899a_49be_8253_369d7c853822.slice/crio-9946567e933aebf50147a88e91a1ba55b2869c9a503620383a61d5ac8847e7cb WatchSource:0}: Error finding container 9946567e933aebf50147a88e91a1ba55b2869c9a503620383a61d5ac8847e7cb: Status 404 returned error can't find the container with id 9946567e933aebf50147a88e91a1ba55b2869c9a503620383a61d5ac8847e7cb Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.580377 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.581121 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.081104522 +0000 UTC m=+150.136003364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.683354 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.683859 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.183840531 +0000 UTC m=+150.238739383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.703035 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.703093 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.720154 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4786l"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.722598 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.735082 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.738233 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.748877 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.750732 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9wm6"] Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.752781 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c"] Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.763941 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb86de852_56af_4d31_b2cf_3c27bd5ff599.slice/crio-6472f28cb566ddf8122006079c6f19fdec91edde484c1f46d55ed9a5b7511a77 WatchSource:0}: Error finding container 6472f28cb566ddf8122006079c6f19fdec91edde484c1f46d55ed9a5b7511a77: Status 404 returned error can't find the container with id 6472f28cb566ddf8122006079c6f19fdec91edde484c1f46d55ed9a5b7511a77 Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.770021 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4"] Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.771853 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ff75676_97f2_4f72_bfbe_4dfd756310d9.slice/crio-eb7c2e69c3cbdfa43096aeccc6ffddbd40a1688535776041c57154b287df9045 WatchSource:0}: Error finding container eb7c2e69c3cbdfa43096aeccc6ffddbd40a1688535776041c57154b287df9045: Status 404 returned error can't find the container with id eb7c2e69c3cbdfa43096aeccc6ffddbd40a1688535776041c57154b287df9045 Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.784467 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.784655 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.284638235 +0000 UTC m=+150.339537077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.784772 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.785019 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.285011019 +0000 UTC m=+150.339909871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.790945 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacfd5bdb_80ee_455d_9c3a_765ab62a7689.slice/crio-e582c85f8983579295d6e9e477dac8480b7b0ed9f8ea24c457db402af980592a WatchSource:0}: Error finding container e582c85f8983579295d6e9e477dac8480b7b0ed9f8ea24c457db402af980592a: Status 404 returned error can't find the container with id e582c85f8983579295d6e9e477dac8480b7b0ed9f8ea24c457db402af980592a Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.799168 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18f2bc53_5f87_46d7_ab1e_2d878b364a46.slice/crio-6c043bfc85a7e342001a993f3edd2481f7768f3cba192c7f27fcc09a59370156 WatchSource:0}: Error finding container 6c043bfc85a7e342001a993f3edd2481f7768f3cba192c7f27fcc09a59370156: Status 404 returned error can't find the container with id 6c043bfc85a7e342001a993f3edd2481f7768f3cba192c7f27fcc09a59370156 Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.801434 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8eefc13_080a_4cf2_928f_0289b856fba1.slice/crio-f0b281a11757b44d70a3d92449138c6cbea9dda8f78f86fe74c7a4fa7692ed97 WatchSource:0}: Error finding container f0b281a11757b44d70a3d92449138c6cbea9dda8f78f86fe74c7a4fa7692ed97: Status 404 returned error can't find the container with id f0b281a11757b44d70a3d92449138c6cbea9dda8f78f86fe74c7a4fa7692ed97 Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.811584 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57ac644c_0670_49bb_b7ce_3d9ecb72fa5a.slice/crio-aaa1e6313cc354ad0aba9c7300c99e256df61561199e2e7718665afe80523bbb WatchSource:0}: Error finding container aaa1e6313cc354ad0aba9c7300c99e256df61561199e2e7718665afe80523bbb: Status 404 returned error can't find the container with id aaa1e6313cc354ad0aba9c7300c99e256df61561199e2e7718665afe80523bbb Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.816801 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdf6f002_40c1_443b_a1fe_89f312da84f2.slice/crio-fc9b3dac3e2023a9552776a8cb1660e1fe410c5329ff31b01ded1ce428c7cf9b WatchSource:0}: Error finding container fc9b3dac3e2023a9552776a8cb1660e1fe410c5329ff31b01ded1ce428c7cf9b: Status 404 returned error can't find the container with id fc9b3dac3e2023a9552776a8cb1660e1fe410c5329ff31b01ded1ce428c7cf9b Oct 04 07:12:12 crc kubenswrapper[4987]: W1004 07:12:12.817190 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7c84c47_7382_42cd_92b0_5e0261774961.slice/crio-b9172c9016bdba3f314f7ba717668a76b4e6fa593af112467ecd185d1a60c59e WatchSource:0}: Error finding container b9172c9016bdba3f314f7ba717668a76b4e6fa593af112467ecd185d1a60c59e: Status 404 returned error can't find the container with id b9172c9016bdba3f314f7ba717668a76b4e6fa593af112467ecd185d1a60c59e Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.892819 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.892967 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.392935659 +0000 UTC m=+150.447834521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.893379 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.893481 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.893522 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.893555 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:12:12 crc kubenswrapper[4987]: E1004 07:12:12.895946 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.395925174 +0000 UTC m=+150.450824016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.897418 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.893617 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.903510 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.904147 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.907208 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:12:12 crc kubenswrapper[4987]: I1004 07:12:12.970857 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" event={"ID":"e3e5bf21-b590-45a3-9891-3e5114d6b25d","Type":"ContainerStarted","Data":"d3b8b3d9cda47e2718be027b3b0a875d036d9fdec08be2f95b5a3d19133028f3"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:12.999665 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" event={"ID":"e2745959-fc00-405c-8c5d-d70e7cb132ec","Type":"ContainerStarted","Data":"1e95987c2bcb4b10e24b5f073657dd8a73e7b53781a302c32e58a0969c6e190b"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.001022 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.001159 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" event={"ID":"d8eefc13-080a-4cf2-928f-0289b856fba1","Type":"ContainerStarted","Data":"f0b281a11757b44d70a3d92449138c6cbea9dda8f78f86fe74c7a4fa7692ed97"} Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.001293 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.501278285 +0000 UTC m=+150.556177137 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.005334 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.007127 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.507115851 +0000 UTC m=+150.562014693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.013367 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" event={"ID":"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649","Type":"ContainerStarted","Data":"7eb5c52fdbfaf6dd7197ba771b3e6cc5324f8344d54d4c09d4b41133f07c1f35"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.017392 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" event={"ID":"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b","Type":"ContainerStarted","Data":"abaea740d3760805a39b646d1267a8e56a5b4be729b964ab72d898ef0923c0bf"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.017429 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" event={"ID":"ffcf2982-cee1-49a6-a2c4-15b23b45ad8b","Type":"ContainerStarted","Data":"539df75483cfa2a0e25db6ed004e3f2309771e5c32d2737b18a8a84727f24908"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.019953 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" event={"ID":"659b7013-28e6-4ccb-b2d6-247c1eff7d4f","Type":"ContainerStarted","Data":"eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.020837 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.023725 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" event={"ID":"7ba6778c-5684-46a3-b7c5-75f26973e3e0","Type":"ContainerStarted","Data":"6fba3423523bdfc2c38cb4a9ad047a7a642da740c6089126fd47b577d47a82ba"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.023769 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" event={"ID":"7ba6778c-5684-46a3-b7c5-75f26973e3e0","Type":"ContainerStarted","Data":"097c2b179c91ba975634f221c874a8df711ea42093effea23469e7277dd081c5"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.028498 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" event={"ID":"5420a718-1d0a-4b72-b57d-a8a4a18e7763","Type":"ContainerStarted","Data":"977640a981b226132b5db893ce87c39752b4d37741c7d0ec2f460e56475388da"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.033723 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" event={"ID":"9021090f-c067-4574-979a-4186a2df4b89","Type":"ContainerStarted","Data":"49d1e3e7bbb84ae6d4f754f34130f52a31cb8f8ccffbd037f508d03ffc854031"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.033957 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" event={"ID":"9021090f-c067-4574-979a-4186a2df4b89","Type":"ContainerStarted","Data":"75452d4de00aa188bf6b1b2db20ae94795f4efc153f1ef66e74da919aa205f94"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.035235 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zzpsj" podStartSLOduration=128.035223149 podStartE2EDuration="2m8.035223149s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.035091764 +0000 UTC m=+150.089990606" watchObservedRunningTime="2025-10-04 07:12:13.035223149 +0000 UTC m=+150.090121991" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.041763 4987 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6hddn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.042396 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.047613 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" event={"ID":"7b91e57c-19f9-49a5-9d92-6be098455d2b","Type":"ContainerStarted","Data":"59ff6fa210ea9e2dcb08f8e5366e447177cb2900cb87d7f73da8b0bdabaab61b"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.050060 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5b448" podStartSLOduration=129.050035683 podStartE2EDuration="2m9.050035683s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:12.989413035 +0000 UTC m=+150.044311877" watchObservedRunningTime="2025-10-04 07:12:13.050035683 +0000 UTC m=+150.104934525" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.063128 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" podStartSLOduration=129.063107879 podStartE2EDuration="2m9.063107879s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.056989482 +0000 UTC m=+150.111888324" watchObservedRunningTime="2025-10-04 07:12:13.063107879 +0000 UTC m=+150.118006721" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.067929 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.070593 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jsh2d" podStartSLOduration=128.070573428 podStartE2EDuration="2m8.070573428s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.070226694 +0000 UTC m=+150.125125536" watchObservedRunningTime="2025-10-04 07:12:13.070573428 +0000 UTC m=+150.125472270" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.075135 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" event={"ID":"3b67efca-532b-437c-b377-2c3bfb7372c0","Type":"ContainerStarted","Data":"af635e3cc24fbecf41d151fceae6aea015d6d85a579b0b762880af44db7b0bfd"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.075177 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" event={"ID":"3b67efca-532b-437c-b377-2c3bfb7372c0","Type":"ContainerStarted","Data":"02173efccf89befffa6afe062eeab186eb7c42d18f73de9b60d5cc13441bb543"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.090049 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" event={"ID":"23705022-817e-4dc7-9303-b2ee99713d42","Type":"ContainerStarted","Data":"22a8b1f98cfa9c499787739ea2789b7947df217df0790341fa2b75f377e757a0"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.092202 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.096794 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.101468 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdbxg" podStartSLOduration=128.101445173 podStartE2EDuration="2m8.101445173s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.092604952 +0000 UTC m=+150.147503794" watchObservedRunningTime="2025-10-04 07:12:13.101445173 +0000 UTC m=+150.156344015" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.106139 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" event={"ID":"b3206d6a-5f6e-4f19-9dac-93b2689734dc","Type":"ContainerStarted","Data":"30737a8c7d18f26335057b68a32aacf050f557157139f6735cbcd6f256a55bbc"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.107004 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.107370 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.109018 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" event={"ID":"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a","Type":"ContainerStarted","Data":"aaa1e6313cc354ad0aba9c7300c99e256df61561199e2e7718665afe80523bbb"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.111301 4987 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-cp8x9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.111339 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" podUID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.111610 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.611582887 +0000 UTC m=+150.666481729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.145862 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-7hkqx" podStartSLOduration=128.145842493 podStartE2EDuration="2m8.145842493s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.122483068 +0000 UTC m=+150.177381910" watchObservedRunningTime="2025-10-04 07:12:13.145842493 +0000 UTC m=+150.200741335" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.147124 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" podStartSLOduration=128.147117153 podStartE2EDuration="2m8.147117153s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.145178827 +0000 UTC m=+150.200077669" watchObservedRunningTime="2025-10-04 07:12:13.147117153 +0000 UTC m=+150.202015995" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.158771 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" event={"ID":"cdf6f002-40c1-443b-a1fe-89f312da84f2","Type":"ContainerStarted","Data":"fc9b3dac3e2023a9552776a8cb1660e1fe410c5329ff31b01ded1ce428c7cf9b"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.162949 4987 generic.go:334] "Generic (PLEG): container finished" podID="262ca319-a044-47bb-b107-1b76e9d9abbb" containerID="cfb46d9758db9ee0456887712129863ad224feda62d1dc3b054794ee9f4d8501" exitCode=0 Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.163155 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt287" event={"ID":"262ca319-a044-47bb-b107-1b76e9d9abbb","Type":"ContainerDied","Data":"cfb46d9758db9ee0456887712129863ad224feda62d1dc3b054794ee9f4d8501"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.180348 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" event={"ID":"acc7a457-28b0-4bd8-98aa-85110a95ae02","Type":"ContainerStarted","Data":"73fc5075a37ec874eddedbee7a952d9e3e336e0a61ccee4376bed56cdba29c78"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.188341 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" event={"ID":"25e66a14-b32b-4aca-9fce-9fc81f5d66f4","Type":"ContainerStarted","Data":"6694d93091173c94198b5f2383b0c5bc68eb26e9db90c701c439587d3ffa84eb"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.190133 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wfp5w" event={"ID":"dc8c0c59-a185-4b91-920e-c53f58d27e61","Type":"ContainerStarted","Data":"cb9d669751994d7c6ee8ddebdffe0e3c5751afba4da99edc0f4ab33f15ce0a94"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.196227 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" event={"ID":"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a","Type":"ContainerStarted","Data":"ce002c28df6e7171a3e98f0495332cceda808168ffbe626d7ecc483bf67bb8ea"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.196258 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" event={"ID":"fc3745bd-0124-4a55-b9b3-a7b27b1aee2a","Type":"ContainerStarted","Data":"1892af93218b85daf7fdab872daa64352e69b13c00dba75480cf864ad5859fe6"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.211389 4987 generic.go:334] "Generic (PLEG): container finished" podID="b475bd2a-4652-40aa-8e9f-0f4a069ad77f" containerID="e465a3316da9d9f36bc84c39cb6417f64d834eaae14ee3ec7a5c2095689f409c" exitCode=0 Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.211783 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" event={"ID":"b475bd2a-4652-40aa-8e9f-0f4a069ad77f","Type":"ContainerDied","Data":"e465a3316da9d9f36bc84c39cb6417f64d834eaae14ee3ec7a5c2095689f409c"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.212553 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.212936 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-rnb5t" podStartSLOduration=129.212925191 podStartE2EDuration="2m9.212925191s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.212200533 +0000 UTC m=+150.267099385" watchObservedRunningTime="2025-10-04 07:12:13.212925191 +0000 UTC m=+150.267824033" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.216389 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.716376635 +0000 UTC m=+150.771275477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.222290 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" event={"ID":"b86de852-56af-4d31-b2cf-3c27bd5ff599","Type":"ContainerStarted","Data":"6472f28cb566ddf8122006079c6f19fdec91edde484c1f46d55ed9a5b7511a77"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.231309 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4786l" event={"ID":"18f2bc53-5f87-46d7-ab1e-2d878b364a46","Type":"ContainerStarted","Data":"6c043bfc85a7e342001a993f3edd2481f7768f3cba192c7f27fcc09a59370156"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.233265 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" event={"ID":"c3ff5bd7-52fa-4f8a-9dcb-2a8941ca1c97","Type":"ContainerStarted","Data":"2431940f0bca3b4060dffeac50acc57cf747a9d1cea65dc9c546f1c2fad7239d"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.255329 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vzg8c" event={"ID":"3c85703d-4a60-4076-b96f-f55fe7483527","Type":"ContainerStarted","Data":"941669f523533ec7f29eb524fe126aced6fe47df74e98859c507296cd8efcfc6"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.259332 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-th4rv" podStartSLOduration=128.259315748 podStartE2EDuration="2m8.259315748s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.235179933 +0000 UTC m=+150.290078775" watchObservedRunningTime="2025-10-04 07:12:13.259315748 +0000 UTC m=+150.314214590" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.260158 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" event={"ID":"f7c84c47-7382-42cd-92b0-5e0261774961","Type":"ContainerStarted","Data":"b9172c9016bdba3f314f7ba717668a76b4e6fa593af112467ecd185d1a60c59e"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.261091 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ggpf" podStartSLOduration=129.261081046 podStartE2EDuration="2m9.261081046s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.257485317 +0000 UTC m=+150.312384159" watchObservedRunningTime="2025-10-04 07:12:13.261081046 +0000 UTC m=+150.315979888" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.267618 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" event={"ID":"7b29632b-7b0b-47ee-b411-87c48881db23","Type":"ContainerStarted","Data":"0ecfd8d6ed380282b2a53d824fc763c6aff02e1cbfda85fb7079437f4dcf260a"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.267675 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" event={"ID":"7b29632b-7b0b-47ee-b411-87c48881db23","Type":"ContainerStarted","Data":"c2c470bacd840c85073b8858b0577c9e1d4099adc1fde92f7b2b5ce5fa6c10db"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.268289 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.272877 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" event={"ID":"acfd5bdb-80ee-455d-9c3a-765ab62a7689","Type":"ContainerStarted","Data":"e582c85f8983579295d6e9e477dac8480b7b0ed9f8ea24c457db402af980592a"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.281833 4987 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-7sbxk container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.282127 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" podUID="7b29632b-7b0b-47ee-b411-87c48881db23" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.287213 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" event={"ID":"d5787789-e288-4011-baeb-17da6439f196","Type":"ContainerStarted","Data":"40e7f6ea521fc14068b92238befe81fea0485766c049c2f5fa08ad14104a5413"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.289135 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bjgbf" event={"ID":"3e1f11da-af03-4bc8-97c8-ad2eac21d489","Type":"ContainerStarted","Data":"deebd684f452980e25fca09a046d94505f7448fcb6780a48e571667dbcf9b571"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.295490 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" event={"ID":"f9a02538-c9f6-41ed-b022-89bafc7d548a","Type":"ContainerStarted","Data":"c030d99304e9a82d62f4a481add5504337d0fc5bcd2e8455d0b726cbd0ef9483"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.295550 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" event={"ID":"f9a02538-c9f6-41ed-b022-89bafc7d548a","Type":"ContainerStarted","Data":"96616252d182588854b7a03cd28c3c794eec2ddd37f0b8881a9063a987f05c20"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.295828 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.297029 4987 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-nv4lt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.297079 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" podUID="f9a02538-c9f6-41ed-b022-89bafc7d548a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.301180 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wfp5w" podStartSLOduration=129.301162688 podStartE2EDuration="2m9.301162688s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.298567198 +0000 UTC m=+150.353466040" watchObservedRunningTime="2025-10-04 07:12:13.301162688 +0000 UTC m=+150.356061550" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.302272 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" event={"ID":"fc1bec83-d97b-403e-9de5-9f06087157ad","Type":"ContainerStarted","Data":"896449eec8acfb41d3357bacac30f5275c19fb2f150f075fdb338c12ba14c4e6"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.303300 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.306813 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nv9l8" event={"ID":"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa","Type":"ContainerStarted","Data":"f4771a5b644cf5abea4339ad184150b413f759b65fcdb8e2f692605fe69b6332"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.315249 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" event={"ID":"d8c9427b-899a-49be-8253-369d7c853822","Type":"ContainerStarted","Data":"9946567e933aebf50147a88e91a1ba55b2869c9a503620383a61d5ac8847e7cb"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.315393 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.316549 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.816528153 +0000 UTC m=+150.871427015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.339741 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" event={"ID":"9498cb9d-681f-4351-b0a2-1606bcc3afdf","Type":"ContainerStarted","Data":"8c8d7561626cfd4542ef7b2c739b61c07b3e0dd245357cb01a1845c6e73d35c4"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.340704 4987 patch_prober.go:28] interesting pod/console-operator-58897d9998-rgcdb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.340775 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" podUID="fc1bec83-d97b-403e-9de5-9f06087157ad" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.403710 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" event={"ID":"5ff75676-97f2-4f72-bfbe-4dfd756310d9","Type":"ContainerStarted","Data":"eb7c2e69c3cbdfa43096aeccc6ffddbd40a1688535776041c57154b287df9045"} Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.403791 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.407208 4987 patch_prober.go:28] interesting pod/downloads-7954f5f757-t2w8v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.407289 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t2w8v" podUID="453b9a94-8a64-477b-97eb-633594674f79" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.410810 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.417774 4987 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tfd59 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.417872 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" podUID="1a5130dd-df25-4642-8180-637b14f8ee4a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.422076 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" podStartSLOduration=128.42204199 podStartE2EDuration="2m8.42204199s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.419852784 +0000 UTC m=+150.474751656" watchObservedRunningTime="2025-10-04 07:12:13.42204199 +0000 UTC m=+150.476940832" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.425524 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xqmwx" podStartSLOduration=129.425500173 podStartE2EDuration="2m9.425500173s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.3714416 +0000 UTC m=+150.426340442" watchObservedRunningTime="2025-10-04 07:12:13.425500173 +0000 UTC m=+150.480399045" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.465785 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.466669 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:13.966641787 +0000 UTC m=+151.021540629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.470994 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" podStartSLOduration=128.470942293 podStartE2EDuration="2m8.470942293s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.458537593 +0000 UTC m=+150.513436435" watchObservedRunningTime="2025-10-04 07:12:13.470942293 +0000 UTC m=+150.525841135" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.545723 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" podStartSLOduration=128.545693058 podStartE2EDuration="2m8.545693058s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.530032742 +0000 UTC m=+150.584931584" watchObservedRunningTime="2025-10-04 07:12:13.545693058 +0000 UTC m=+150.600591900" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.557111 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vzg8c" podStartSLOduration=6.557085469 podStartE2EDuration="6.557085469s" podCreationTimestamp="2025-10-04 07:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.55116913 +0000 UTC m=+150.606067972" watchObservedRunningTime="2025-10-04 07:12:13.557085469 +0000 UTC m=+150.611984311" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.572378 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.578258 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" podStartSLOduration=129.578236238 podStartE2EDuration="2m9.578236238s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.577666036 +0000 UTC m=+150.632564878" watchObservedRunningTime="2025-10-04 07:12:13.578236238 +0000 UTC m=+150.633135080" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.578216 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.078183146 +0000 UTC m=+151.133081978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.663567 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bjgbf" podStartSLOduration=128.663548302 podStartE2EDuration="2m8.663548302s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.636173702 +0000 UTC m=+150.691072554" watchObservedRunningTime="2025-10-04 07:12:13.663548302 +0000 UTC m=+150.718447144" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.681330 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.681872 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.181851371 +0000 UTC m=+151.236750223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.727667 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" podStartSLOduration=129.727645924 podStartE2EDuration="2m9.727645924s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.664749969 +0000 UTC m=+150.719648811" watchObservedRunningTime="2025-10-04 07:12:13.727645924 +0000 UTC m=+150.782544766" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.741063 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-9mxgt" podStartSLOduration=129.741045204 podStartE2EDuration="2m9.741045204s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.739584067 +0000 UTC m=+150.794482909" watchObservedRunningTime="2025-10-04 07:12:13.741045204 +0000 UTC m=+150.795944046" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.769511 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-t2w8v" podStartSLOduration=129.769485424 podStartE2EDuration="2m9.769485424s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:13.768304529 +0000 UTC m=+150.823203371" watchObservedRunningTime="2025-10-04 07:12:13.769485424 +0000 UTC m=+150.824384266" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.793750 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.794110 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.294067516 +0000 UTC m=+151.348966358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.794332 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.794690 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.29467593 +0000 UTC m=+151.349574772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.827177 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.830050 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.830162 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.895312 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.895504 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.395468024 +0000 UTC m=+151.450366866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: E1004 07:12:13.896253 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.396242404 +0000 UTC m=+151.451141246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:13 crc kubenswrapper[4987]: I1004 07:12:13.908702 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.025358 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.026163 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.526136204 +0000 UTC m=+151.581035046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.128866 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.129320 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.62930628 +0000 UTC m=+151.684205122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.231239 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.232265 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.732238776 +0000 UTC m=+151.787137628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.335561 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.336030 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.836015895 +0000 UTC m=+151.890914737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.447227 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.449344 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:14.949314452 +0000 UTC m=+152.004213304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.468697 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" event={"ID":"9498cb9d-681f-4351-b0a2-1606bcc3afdf","Type":"ContainerStarted","Data":"69281df6497d78ade66328d9601c80533e26e725756ef89937b0dab5f6997083"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.503954 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" event={"ID":"5ff75676-97f2-4f72-bfbe-4dfd756310d9","Type":"ContainerStarted","Data":"b8bf41f5782e06af6616234c28078d2a814d9f501832c27ab81c122d2dc3294d"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.504013 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.507742 4987 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vw86c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.507833 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" podUID="5ff75676-97f2-4f72-bfbe-4dfd756310d9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.509451 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" event={"ID":"23705022-817e-4dc7-9303-b2ee99713d42","Type":"ContainerStarted","Data":"6d9a344468d684523268a64efe6cde6a5e472529fd76b8462870fbf1d43738dd"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.513253 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4786l" event={"ID":"18f2bc53-5f87-46d7-ab1e-2d878b364a46","Type":"ContainerStarted","Data":"29b3064d2be0814bfe12b692015f5183ad226f4420a564a0143c73e8e93c6809"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.521055 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3da87b2cf82c1a15c1a7e9498df0611bab3fa818a85f3d8c6da23627287b105c"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.522308 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" event={"ID":"cdf6f002-40c1-443b-a1fe-89f312da84f2","Type":"ContainerStarted","Data":"6e4e95919e0fd5a9942fbc0514e20a8559056f11e440541eecbfc76b9325e2a1"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.524585 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" event={"ID":"acfd5bdb-80ee-455d-9c3a-765ab62a7689","Type":"ContainerStarted","Data":"b5f06c3bd583ded96bacec3015bc0f1a082d6756eb7dc8f6caf70615580f1b21"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.524667 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" event={"ID":"acfd5bdb-80ee-455d-9c3a-765ab62a7689","Type":"ContainerStarted","Data":"ce18cf5c8701d21345e8a2d572a2ad674c3a092e317fc14660395538185c18f7"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.528974 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" event={"ID":"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649","Type":"ContainerStarted","Data":"06006a63ca6941e721f3852c71d1e6347d189e510954757ad89c40cbb815f945"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.529024 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" event={"ID":"35efc8ac-5b7d-4fbc-828d-cf84bb7d8649","Type":"ContainerStarted","Data":"83d17fd78c9cf3731ccef786addace84fa15733c1cf30533d1caec3a878837e4"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.530402 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nv9l8" event={"ID":"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa","Type":"ContainerStarted","Data":"dec1c4e14694b943b48a3c0059343570dfdad84c57096a8942f676971c2bee37"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.531034 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"64e586a3a091c6cbd93519a512c7541808fd46f6dc563e6a1e12bdc2eebef909"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.532086 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" event={"ID":"e2745959-fc00-405c-8c5d-d70e7cb132ec","Type":"ContainerStarted","Data":"cfe3e39c44f17dd390df39bd37fb8afe14bc698a452f0647d920a2774fa84ba8"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.533360 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" event={"ID":"b86de852-56af-4d31-b2cf-3c27bd5ff599","Type":"ContainerStarted","Data":"c31c4424e21e89af214a8896edab6ec227c72a32c0c55289a6b169e26bf53da2"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.541576 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" event={"ID":"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a","Type":"ContainerStarted","Data":"b7acd789d89c4ef81a64ca88670dbbbce5fa6aaca5b5c3b741b29cfb477b1c5b"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.542355 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.547122 4987 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j9wm6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.547208 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.550691 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" event={"ID":"9021090f-c067-4574-979a-4186a2df4b89","Type":"ContainerStarted","Data":"fdab3417af1599b5b57bdbe5ab84e9c7323400731bfe0c9ec4dee9c7c0c7c5c3"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.550830 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.552324 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.052308651 +0000 UTC m=+152.107207493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.561787 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1338a15f26c16ace2bfa05f00ac90bc619c3870abf42052105a77c6b6fb41f57"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.567783 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" event={"ID":"f7c84c47-7382-42cd-92b0-5e0261774961","Type":"ContainerStarted","Data":"9fbc4f193fb70b1eceda96c3cbeb5d606338a1d591f9ae75cfe322f1511b6eda"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.569671 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" event={"ID":"b475bd2a-4652-40aa-8e9f-0f4a069ad77f","Type":"ContainerStarted","Data":"bfd558908cb4bd9d83870ea29ec3eba21bc14324d8eb32c4c693c8fb79f7ad37"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.570167 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.571895 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" event={"ID":"d8eefc13-080a-4cf2-928f-0289b856fba1","Type":"ContainerStarted","Data":"8efe7215d3cea524f7077aedd62d3fde7d6115dec3a59f37c050702932830cdd"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.571926 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" event={"ID":"d8eefc13-080a-4cf2-928f-0289b856fba1","Type":"ContainerStarted","Data":"a1b00c23b73331c8dbbb900b50a53e5994fbbd66f5d67028bc3a6413be348b17"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.574112 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f49kw" event={"ID":"d5787789-e288-4011-baeb-17da6439f196","Type":"ContainerStarted","Data":"e529c8ed3a8337f84ce4d44b758b0b3800f78a83fc15c767eba73d4efc53ea7f"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.577916 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" event={"ID":"7b91e57c-19f9-49a5-9d92-6be098455d2b","Type":"ContainerStarted","Data":"acb8994ee2c841c8bc6ab600d2ae325e7d1add72e29635ea56788ac05b70851d"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.578038 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" event={"ID":"7b91e57c-19f9-49a5-9d92-6be098455d2b","Type":"ContainerStarted","Data":"b10c437af8dee30baa0f903c609e43e111904d4ece8d496e10152331613e130d"} Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.578550 4987 patch_prober.go:28] interesting pod/downloads-7954f5f757-t2w8v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.578666 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t2w8v" podUID="453b9a94-8a64-477b-97eb-633594674f79" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.578665 4987 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6hddn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.578800 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.579837 4987 patch_prober.go:28] interesting pod/console-operator-58897d9998-rgcdb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.579904 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" podUID="fc1bec83-d97b-403e-9de5-9f06087157ad" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.581030 4987 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-7sbxk container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.581071 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" podUID="7b29632b-7b0b-47ee-b411-87c48881db23" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.582159 4987 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-nv4lt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.582207 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" podUID="f9a02538-c9f6-41ed-b022-89bafc7d548a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.595298 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.651443 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.651755 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.151737551 +0000 UTC m=+152.206636393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.655935 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.665067 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.165053917 +0000 UTC m=+152.219952759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.699111 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-c4hjg" podStartSLOduration=129.699093426 podStartE2EDuration="2m9.699093426s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.699044444 +0000 UTC m=+151.753943286" watchObservedRunningTime="2025-10-04 07:12:14.699093426 +0000 UTC m=+151.753992268" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.742014 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-b5z4v" podStartSLOduration=129.741996497 podStartE2EDuration="2m9.741996497s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.739029582 +0000 UTC m=+151.793928424" watchObservedRunningTime="2025-10-04 07:12:14.741996497 +0000 UTC m=+151.796895339" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.772215 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" podStartSLOduration=129.772198347 podStartE2EDuration="2m9.772198347s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.770265892 +0000 UTC m=+151.825164734" watchObservedRunningTime="2025-10-04 07:12:14.772198347 +0000 UTC m=+151.827097189" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.774486 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.782879 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.282854289 +0000 UTC m=+152.337753131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.831235 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:14 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:14 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:14 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.834228 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.841852 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" podStartSLOduration=129.841830693 podStartE2EDuration="2m9.841830693s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.825573874 +0000 UTC m=+151.880472716" watchObservedRunningTime="2025-10-04 07:12:14.841830693 +0000 UTC m=+151.896729535" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.884530 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.884904 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.384890741 +0000 UTC m=+152.439789583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.909715 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7z9pt" podStartSLOduration=129.909697062 podStartE2EDuration="2m9.909697062s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.909042206 +0000 UTC m=+151.963941058" watchObservedRunningTime="2025-10-04 07:12:14.909697062 +0000 UTC m=+151.964595904" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.910567 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" podStartSLOduration=130.910561945 podStartE2EDuration="2m10.910561945s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.880105116 +0000 UTC m=+151.935003948" watchObservedRunningTime="2025-10-04 07:12:14.910561945 +0000 UTC m=+151.965460787" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.948108 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-v9cn9" podStartSLOduration=129.948083818 podStartE2EDuration="2m9.948083818s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.943133596 +0000 UTC m=+151.998032448" watchObservedRunningTime="2025-10-04 07:12:14.948083818 +0000 UTC m=+152.002982660" Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.985706 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.986170 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.486121291 +0000 UTC m=+152.541020133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:14 crc kubenswrapper[4987]: I1004 07:12:14.986476 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:14 crc kubenswrapper[4987]: E1004 07:12:14.987071 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.487063357 +0000 UTC m=+152.541962199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.008590 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c2hns" podStartSLOduration=130.00856379 podStartE2EDuration="2m10.00856379s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:14.97498705 +0000 UTC m=+152.029885892" watchObservedRunningTime="2025-10-04 07:12:15.00856379 +0000 UTC m=+152.063462632" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.008820 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bbx6q" podStartSLOduration=131.00881582 podStartE2EDuration="2m11.00881582s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.006717548 +0000 UTC m=+152.061616390" watchObservedRunningTime="2025-10-04 07:12:15.00881582 +0000 UTC m=+152.063714662" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.034999 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96m8f" podStartSLOduration=131.034973983 podStartE2EDuration="2m11.034973983s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.031273569 +0000 UTC m=+152.086172431" watchObservedRunningTime="2025-10-04 07:12:15.034973983 +0000 UTC m=+152.089872825" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.076447 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qv6pn" podStartSLOduration=130.076421278 podStartE2EDuration="2m10.076421278s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.074322067 +0000 UTC m=+152.129220909" watchObservedRunningTime="2025-10-04 07:12:15.076421278 +0000 UTC m=+152.131320120" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.090692 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.091246 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.591203361 +0000 UTC m=+152.646102203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.105175 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vtxt4" podStartSLOduration=130.105154781 podStartE2EDuration="2m10.105154781s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.104583229 +0000 UTC m=+152.159482071" watchObservedRunningTime="2025-10-04 07:12:15.105154781 +0000 UTC m=+152.160053623" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.122731 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" podStartSLOduration=131.12270995 podStartE2EDuration="2m11.12270995s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.120015877 +0000 UTC m=+152.174914719" watchObservedRunningTime="2025-10-04 07:12:15.12270995 +0000 UTC m=+152.177608792" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.148648 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4786l" podStartSLOduration=8.148612744 podStartE2EDuration="8.148612744s" podCreationTimestamp="2025-10-04 07:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.143643271 +0000 UTC m=+152.198542113" watchObservedRunningTime="2025-10-04 07:12:15.148612744 +0000 UTC m=+152.203511606" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.192261 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.192757 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.692734132 +0000 UTC m=+152.747632984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.293550 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.293738 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.793697003 +0000 UTC m=+152.848595845 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.294134 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.294463 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.794452031 +0000 UTC m=+152.849350873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.344848 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.395694 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.396028 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.896013625 +0000 UTC m=+152.950912467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.497386 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.497751 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:15.997740434 +0000 UTC m=+153.052639276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.584121 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e851db37081b1d4e3c5038561a16e6226849d1a2d9cc0886e298b24f4eacbe0a"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.586588 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"caf493fb10658e7858ee634fd26cf2bdd88dd3bc7b0006806bf51106bd9be89e"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.589196 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt287" event={"ID":"262ca319-a044-47bb-b107-1b76e9d9abbb","Type":"ContainerStarted","Data":"74396495007a15e0bf1bf33525d6da161fd101f24c09f589eee92a83ca3b685d"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.589221 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-wt287" event={"ID":"262ca319-a044-47bb-b107-1b76e9d9abbb","Type":"ContainerStarted","Data":"675a620eaa07a85935f7c64004087dc91be3fcc7f906bdf4b425c31bc85d11f5"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.591182 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" event={"ID":"8e6f14f4-f9b0-438a-a4b4-290b3beb7814","Type":"ContainerStarted","Data":"57c5fe9ce01951a758b64c7ac4b32c07de5cb61cddc5f3b72a10a9a430af8c29"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.592887 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nv9l8" event={"ID":"bbe932ef-c8f2-4eba-a23b-226aaf51ceaa","Type":"ContainerStarted","Data":"704586534ffca8312cbb7ab204fd2dce5d98fd5e6165e5591cf22534ec1ed6df"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.593210 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.595148 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d346250dabdb076b27ff2c0efa1543473a43533c031b409d55c80f25a6f27a79"} Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.595194 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.596162 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.601085 4987 patch_prober.go:28] interesting pod/console-operator-58897d9998-rgcdb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.601136 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.601088 4987 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vw86c container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.601160 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" podUID="fc1bec83-d97b-403e-9de5-9f06087157ad" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.601210 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" podUID="5ff75676-97f2-4f72-bfbe-4dfd756310d9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.601379 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.101366818 +0000 UTC m=+153.156265660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.602268 4987 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-j9wm6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.602308 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.703249 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.705873 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.205856054 +0000 UTC m=+153.260754896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.751604 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-wt287" podStartSLOduration=131.751568804 podStartE2EDuration="2m11.751568804s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.737068063 +0000 UTC m=+152.791966905" watchObservedRunningTime="2025-10-04 07:12:15.751568804 +0000 UTC m=+152.806467646" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.805374 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.805582 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.305551095 +0000 UTC m=+153.360449937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.806121 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.806443 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.306433469 +0000 UTC m=+153.361332481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.812898 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" podStartSLOduration=130.812879099 podStartE2EDuration="2m10.812879099s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.808578522 +0000 UTC m=+152.863477364" watchObservedRunningTime="2025-10-04 07:12:15.812879099 +0000 UTC m=+152.867777941" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.822700 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:15 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:15 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:15 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.823136 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.843130 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-nv9l8" podStartSLOduration=8.84310993 podStartE2EDuration="8.84310993s" podCreationTimestamp="2025-10-04 07:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.842012687 +0000 UTC m=+152.896911529" watchObservedRunningTime="2025-10-04 07:12:15.84310993 +0000 UTC m=+152.898008772" Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.906894 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:15 crc kubenswrapper[4987]: E1004 07:12:15.907169 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.40715267 +0000 UTC m=+153.462051512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:15 crc kubenswrapper[4987]: I1004 07:12:15.966303 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" podStartSLOduration=130.96628756 podStartE2EDuration="2m10.96628756s" podCreationTimestamp="2025-10-04 07:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:15.891869148 +0000 UTC m=+152.946767980" watchObservedRunningTime="2025-10-04 07:12:15.96628756 +0000 UTC m=+153.021186402" Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.009240 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.009720 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.509693611 +0000 UTC m=+153.564592623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.110465 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.110669 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.61064346 +0000 UTC m=+153.665542302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.111048 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.111418 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.61141011 +0000 UTC m=+153.666308952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.212325 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.212527 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.712501015 +0000 UTC m=+153.767399867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.212848 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.213278 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.713239854 +0000 UTC m=+153.768138696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.226324 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.313725 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.314097 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.814069208 +0000 UTC m=+153.868968050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.415519 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.415924 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:16.915904602 +0000 UTC m=+153.970803494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.517366 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.517753 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.017724395 +0000 UTC m=+154.072623237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.600829 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" event={"ID":"d8c9427b-899a-49be-8253-369d7c853822","Type":"ContainerStarted","Data":"d53504ba62e296ac684ad0e573f47dedb4e9e1e54531f826df67e72b1bfe0f09"} Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.618707 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.619123 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.119106011 +0000 UTC m=+154.174004853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.719551 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.719728 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.219702147 +0000 UTC m=+154.274600989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.720096 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.721652 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.221616731 +0000 UTC m=+154.276515573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.826338 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.826521 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.326491653 +0000 UTC m=+154.381390495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.826651 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.826968 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.32695469 +0000 UTC m=+154.381853542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.831464 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:16 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:16 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:16 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.831521 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.927581 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.927732 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.427709783 +0000 UTC m=+154.482608625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:16 crc kubenswrapper[4987]: I1004 07:12:16.927879 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:16 crc kubenswrapper[4987]: E1004 07:12:16.929084 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.429071146 +0000 UTC m=+154.483969988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.029187 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.029388 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.529359159 +0000 UTC m=+154.584258001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.029442 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.029788 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.529774106 +0000 UTC m=+154.584672948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.130985 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.131380 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.631354999 +0000 UTC m=+154.686253841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.131418 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.131738 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.631729904 +0000 UTC m=+154.686628746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.232485 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.232680 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.732654162 +0000 UTC m=+154.787552994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.232778 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.233094 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.733084899 +0000 UTC m=+154.787983821 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.333989 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.334191 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.834159774 +0000 UTC m=+154.889058616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.435981 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.436251 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:17.936239196 +0000 UTC m=+154.991138038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.493375 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8lddn"] Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.494766 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.505111 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.549691 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8lddn"] Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.550687 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.551179 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.051163997 +0000 UTC m=+155.106062839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.624670 4987 generic.go:334] "Generic (PLEG): container finished" podID="cdf6f002-40c1-443b-a1fe-89f312da84f2" containerID="6e4e95919e0fd5a9942fbc0514e20a8559056f11e440541eecbfc76b9325e2a1" exitCode=0 Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.624756 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" event={"ID":"cdf6f002-40c1-443b-a1fe-89f312da84f2","Type":"ContainerDied","Data":"6e4e95919e0fd5a9942fbc0514e20a8559056f11e440541eecbfc76b9325e2a1"} Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.645273 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" event={"ID":"d8c9427b-899a-49be-8253-369d7c853822","Type":"ContainerStarted","Data":"b4098353d58c4606987ce2725b60cf3aeca38cc904798aa89a96c0625636709d"} Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.652361 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-utilities\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.652408 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-catalog-content\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.652444 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.652473 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9t2\" (UniqueName: \"kubernetes.io/projected/c5530de4-963e-4282-9e46-5ead5330121f-kube-api-access-4x9t2\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.652846 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.152832924 +0000 UTC m=+155.207731766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.674965 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-98fbt"] Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.685611 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.689225 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.693542 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98fbt"] Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.753389 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.753734 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-utilities\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.753791 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-catalog-content\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.753891 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9t2\" (UniqueName: \"kubernetes.io/projected/c5530de4-963e-4282-9e46-5ead5330121f-kube-api-access-4x9t2\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.754460 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-catalog-content\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.754677 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.254653148 +0000 UTC m=+155.309551990 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.755146 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-utilities\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.814740 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9t2\" (UniqueName: \"kubernetes.io/projected/c5530de4-963e-4282-9e46-5ead5330121f-kube-api-access-4x9t2\") pod \"certified-operators-8lddn\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.824485 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.828366 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:17 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:17 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:17 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.828426 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.856355 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwldq\" (UniqueName: \"kubernetes.io/projected/2841f4fc-7a32-4d8a-91a7-f088396830f3-kube-api-access-cwldq\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.856667 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.856793 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-catalog-content\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.856988 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-utilities\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.857357 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.357344735 +0000 UTC m=+155.412243577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.879839 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p9k8v"] Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.881045 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.899888 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9k8v"] Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.958268 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.958589 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-utilities\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.958619 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwldq\" (UniqueName: \"kubernetes.io/projected/2841f4fc-7a32-4d8a-91a7-f088396830f3-kube-api-access-cwldq\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.958664 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-catalog-content\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.959114 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-catalog-content\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: E1004 07:12:17.959187 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.459172368 +0000 UTC m=+155.514071210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.959399 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-utilities\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:17 crc kubenswrapper[4987]: I1004 07:12:17.982304 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwldq\" (UniqueName: \"kubernetes.io/projected/2841f4fc-7a32-4d8a-91a7-f088396830f3-kube-api-access-cwldq\") pod \"community-operators-98fbt\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.009014 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.060375 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.060416 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-utilities\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.060444 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-catalog-content\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.060464 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkgqs\" (UniqueName: \"kubernetes.io/projected/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-kube-api-access-vkgqs\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.060792 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.560777463 +0000 UTC m=+155.615676305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.078195 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gls68"] Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.079823 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.098231 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gls68"] Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.164044 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.164560 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-catalog-content\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.164593 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkgqs\" (UniqueName: \"kubernetes.io/projected/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-kube-api-access-vkgqs\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.164710 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-utilities\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.165148 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-utilities\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.165395 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-catalog-content\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.165796 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.665779279 +0000 UTC m=+155.720678121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.201829 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkgqs\" (UniqueName: \"kubernetes.io/projected/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-kube-api-access-vkgqs\") pod \"certified-operators-p9k8v\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.228497 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.282732 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-utilities\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.282797 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-catalog-content\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.282861 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8zgq\" (UniqueName: \"kubernetes.io/projected/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-kube-api-access-p8zgq\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.282901 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.283211 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.783200036 +0000 UTC m=+155.838098878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.361076 4987 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.386548 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.387114 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-catalog-content\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.387163 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8zgq\" (UniqueName: \"kubernetes.io/projected/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-kube-api-access-p8zgq\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.387237 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-utilities\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.387728 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-utilities\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.387813 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.887795617 +0000 UTC m=+155.942694459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.388066 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-catalog-content\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.410827 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8zgq\" (UniqueName: \"kubernetes.io/projected/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-kube-api-access-p8zgq\") pod \"community-operators-gls68\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.427725 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8lddn"] Oct 04 07:12:18 crc kubenswrapper[4987]: W1004 07:12:18.438189 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5530de4_963e_4282_9e46_5ead5330121f.slice/crio-6235be29ed2a2248e7dba967f6f27c98a26ac3aed22a091b4326e4e1a7d62201 WatchSource:0}: Error finding container 6235be29ed2a2248e7dba967f6f27c98a26ac3aed22a091b4326e4e1a7d62201: Status 404 returned error can't find the container with id 6235be29ed2a2248e7dba967f6f27c98a26ac3aed22a091b4326e4e1a7d62201 Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.488421 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.488777 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:18.988766467 +0000 UTC m=+156.043665309 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.523691 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-98fbt"] Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.598980 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.599809 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.099791987 +0000 UTC m=+156.154690829 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.661586 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerStarted","Data":"d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e"} Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.661640 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerStarted","Data":"6235be29ed2a2248e7dba967f6f27c98a26ac3aed22a091b4326e4e1a7d62201"} Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.673083 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" event={"ID":"d8c9427b-899a-49be-8253-369d7c853822","Type":"ContainerStarted","Data":"aa38e948c4bee58f2d14fd44b43f1a99bf9d07146d7e8cc3b191e9e1a779aae8"} Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.673148 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" event={"ID":"d8c9427b-899a-49be-8253-369d7c853822","Type":"ContainerStarted","Data":"a9ccf00c1931ba63600a693a70820853753932b2642b6bbe95a6d40e11c424e8"} Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.676311 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbt" event={"ID":"2841f4fc-7a32-4d8a-91a7-f088396830f3","Type":"ContainerStarted","Data":"22b2b5f60d4254cb1b4fb720bbfc834b5766b47f54b400656bdb82664874dac4"} Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.700172 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.701574 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.702336 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.202317747 +0000 UTC m=+156.257216769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.731498 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pdb9r" podStartSLOduration=11.731472847 podStartE2EDuration="11.731472847s" podCreationTimestamp="2025-10-04 07:12:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:18.729054133 +0000 UTC m=+155.783952975" watchObservedRunningTime="2025-10-04 07:12:18.731472847 +0000 UTC m=+155.786371689" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.802837 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.803096 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.303069169 +0000 UTC m=+156.357968021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.803640 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.803929 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.303917213 +0000 UTC m=+156.358816055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.824017 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:18 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:18 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:18 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.824062 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.895994 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9k8v"] Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.904224 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:18 crc kubenswrapper[4987]: E1004 07:12:18.904447 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.404433285 +0000 UTC m=+156.459332127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.927476 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.928665 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.932331 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.932736 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.940744 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.962827 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:18 crc kubenswrapper[4987]: I1004 07:12:18.996282 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gls68"] Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.004698 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cdf6f002-40c1-443b-a1fe-89f312da84f2-secret-volume\") pod \"cdf6f002-40c1-443b-a1fe-89f312da84f2\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.004895 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdf6f002-40c1-443b-a1fe-89f312da84f2-config-volume\") pod \"cdf6f002-40c1-443b-a1fe-89f312da84f2\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.004948 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv5fd\" (UniqueName: \"kubernetes.io/projected/cdf6f002-40c1-443b-a1fe-89f312da84f2-kube-api-access-pv5fd\") pod \"cdf6f002-40c1-443b-a1fe-89f312da84f2\" (UID: \"cdf6f002-40c1-443b-a1fe-89f312da84f2\") " Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.005083 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.005124 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.005153 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:19 crc kubenswrapper[4987]: E1004 07:12:19.005435 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.505419626 +0000 UTC m=+156.560318468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.005752 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdf6f002-40c1-443b-a1fe-89f312da84f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "cdf6f002-40c1-443b-a1fe-89f312da84f2" (UID: "cdf6f002-40c1-443b-a1fe-89f312da84f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.015291 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf6f002-40c1-443b-a1fe-89f312da84f2-kube-api-access-pv5fd" (OuterVolumeSpecName: "kube-api-access-pv5fd") pod "cdf6f002-40c1-443b-a1fe-89f312da84f2" (UID: "cdf6f002-40c1-443b-a1fe-89f312da84f2"). InnerVolumeSpecName "kube-api-access-pv5fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.018015 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf6f002-40c1-443b-a1fe-89f312da84f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cdf6f002-40c1-443b-a1fe-89f312da84f2" (UID: "cdf6f002-40c1-443b-a1fe-89f312da84f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.105728 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:19 crc kubenswrapper[4987]: E1004 07:12:19.107506 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.606048073 +0000 UTC m=+156.660946915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.107648 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.107880 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.108019 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.108332 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv5fd\" (UniqueName: \"kubernetes.io/projected/cdf6f002-40c1-443b-a1fe-89f312da84f2-kube-api-access-pv5fd\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.108417 4987 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cdf6f002-40c1-443b-a1fe-89f312da84f2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.108478 4987 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdf6f002-40c1-443b-a1fe-89f312da84f2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:19 crc kubenswrapper[4987]: E1004 07:12:19.108909 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.608898183 +0000 UTC m=+156.663797025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mrxhg" (UID: "001408b5-e454-432d-93dc-3a2dc497941f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.109030 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.132400 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.209404 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:19 crc kubenswrapper[4987]: E1004 07:12:19.209671 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 07:12:19.709655315 +0000 UTC m=+156.764554157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.225153 4987 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T07:12:18.361105143Z","Handler":null,"Name":""} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.227273 4987 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.227312 4987 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.257474 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.310487 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.312513 4987 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.312539 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.341880 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mrxhg\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.411260 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.420175 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.424302 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.454426 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.468234 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m92l5"] Oct 04 07:12:19 crc kubenswrapper[4987]: E1004 07:12:19.469077 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf6f002-40c1-443b-a1fe-89f312da84f2" containerName="collect-profiles" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.469092 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf6f002-40c1-443b-a1fe-89f312da84f2" containerName="collect-profiles" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.469218 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf6f002-40c1-443b-a1fe-89f312da84f2" containerName="collect-profiles" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.475381 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.477264 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.481249 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m92l5"] Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.512812 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-catalog-content\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.512881 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-utilities\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.512969 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqcbx\" (UniqueName: \"kubernetes.io/projected/b0640587-71f4-44e4-a6a5-5dc267d57844-kube-api-access-lqcbx\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.609337 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-f4gb9" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.614304 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-catalog-content\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.614377 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-utilities\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.614405 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqcbx\" (UniqueName: \"kubernetes.io/projected/b0640587-71f4-44e4-a6a5-5dc267d57844-kube-api-access-lqcbx\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.615677 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-catalog-content\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.616151 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-utilities\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.636250 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqcbx\" (UniqueName: \"kubernetes.io/projected/b0640587-71f4-44e4-a6a5-5dc267d57844-kube-api-access-lqcbx\") pod \"redhat-marketplace-m92l5\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.721652 4987 generic.go:334] "Generic (PLEG): container finished" podID="c5530de4-963e-4282-9e46-5ead5330121f" containerID="d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e" exitCode=0 Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.721760 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerDied","Data":"d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.723968 4987 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.724681 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d5f0ecd-6403-4646-9197-3826ef18b8f6","Type":"ContainerStarted","Data":"58d03bd3b00c52f9577a5c5ddcac5e9e0499b07cc6edc84c46161626b38a60ab"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.726943 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.733849 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326020-h7n99" event={"ID":"cdf6f002-40c1-443b-a1fe-89f312da84f2","Type":"ContainerDied","Data":"fc9b3dac3e2023a9552776a8cb1660e1fe410c5329ff31b01ded1ce428c7cf9b"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.736020 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc9b3dac3e2023a9552776a8cb1660e1fe410c5329ff31b01ded1ce428c7cf9b" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.737177 4987 generic.go:334] "Generic (PLEG): container finished" podID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerID="8ba6f84942c6cabd9e1b63cb1a11feadc3a3fda213dd8cb39e381cc888e43a96" exitCode=0 Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.737236 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9k8v" event={"ID":"36b7fa18-bb86-4094-b3a8-2e8323fa2a05","Type":"ContainerDied","Data":"8ba6f84942c6cabd9e1b63cb1a11feadc3a3fda213dd8cb39e381cc888e43a96"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.737260 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9k8v" event={"ID":"36b7fa18-bb86-4094-b3a8-2e8323fa2a05","Type":"ContainerStarted","Data":"7fbdae595c56270c975ab4b2151e36a7e1658c714323d85e94c90713eef04430"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.746979 4987 generic.go:334] "Generic (PLEG): container finished" podID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerID="0f0e301c86a98f19a06f0d9c527b53ecb2cc3688e836bd33038925bb79663c1e" exitCode=0 Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.747173 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbt" event={"ID":"2841f4fc-7a32-4d8a-91a7-f088396830f3","Type":"ContainerDied","Data":"0f0e301c86a98f19a06f0d9c527b53ecb2cc3688e836bd33038925bb79663c1e"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.750020 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mrxhg"] Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.752338 4987 generic.go:334] "Generic (PLEG): container finished" podID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerID="cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865" exitCode=0 Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.752520 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gls68" event={"ID":"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af","Type":"ContainerDied","Data":"cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865"} Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.752555 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gls68" event={"ID":"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af","Type":"ContainerStarted","Data":"5e7b9148a9f2831edce4b7a1c287269404ec1783d4dae8c225a3850f7bdb072d"} Oct 04 07:12:19 crc kubenswrapper[4987]: W1004 07:12:19.758274 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod001408b5_e454_432d_93dc_3a2dc497941f.slice/crio-4c55978594c509d2aa4b298fdc59a5cdd9768cc904a52ab9627ee016674e5012 WatchSource:0}: Error finding container 4c55978594c509d2aa4b298fdc59a5cdd9768cc904a52ab9627ee016674e5012: Status 404 returned error can't find the container with id 4c55978594c509d2aa4b298fdc59a5cdd9768cc904a52ab9627ee016674e5012 Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.820869 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:19 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:19 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:19 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.820935 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.833745 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.865362 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6fdjr"] Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.866604 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.872105 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fdjr"] Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.912279 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.912333 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.918125 4987 patch_prober.go:28] interesting pod/apiserver-76f77b778f-wt287 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]log ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]etcd ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/max-in-flight-filter ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 04 07:12:19 crc kubenswrapper[4987]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 04 07:12:19 crc kubenswrapper[4987]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/project.openshift.io-projectcache ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/openshift.io-startinformers ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 04 07:12:19 crc kubenswrapper[4987]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 07:12:19 crc kubenswrapper[4987]: livez check failed Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.918169 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-wt287" podUID="262ca319-a044-47bb-b107-1b76e9d9abbb" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.918837 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-utilities\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.918999 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-catalog-content\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.919047 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59kfc\" (UniqueName: \"kubernetes.io/projected/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-kube-api-access-59kfc\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.978889 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.979274 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:19 crc kubenswrapper[4987]: I1004 07:12:19.984579 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.020741 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59kfc\" (UniqueName: \"kubernetes.io/projected/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-kube-api-access-59kfc\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.020866 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-utilities\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.020973 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-catalog-content\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.021582 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-catalog-content\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.021883 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-utilities\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.047366 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59kfc\" (UniqueName: \"kubernetes.io/projected/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-kube-api-access-59kfc\") pod \"redhat-marketplace-6fdjr\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.066384 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.079253 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m92l5"] Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.117337 4987 patch_prober.go:28] interesting pod/downloads-7954f5f757-t2w8v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.117366 4987 patch_prober.go:28] interesting pod/downloads-7954f5f757-t2w8v container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.117403 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-t2w8v" podUID="453b9a94-8a64-477b-97eb-633594674f79" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.117403 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t2w8v" podUID="453b9a94-8a64-477b-97eb-633594674f79" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.183678 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.342389 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.342798 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.343999 4987 patch_prober.go:28] interesting pod/console-f9d7485db-wfp5w container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.344049 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wfp5w" podUID="dc8c0c59-a185-4b91-920e-c53f58d27e61" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.383020 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fdjr"] Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.416275 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rgcdb" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.778096 4987 generic.go:334] "Generic (PLEG): container finished" podID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerID="7898568fbe78517afc4139ca937320d296e058a896b7293f5b3e5e73f6b75ca0" exitCode=0 Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.778502 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fdjr" event={"ID":"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d","Type":"ContainerDied","Data":"7898568fbe78517afc4139ca937320d296e058a896b7293f5b3e5e73f6b75ca0"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.778546 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fdjr" event={"ID":"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d","Type":"ContainerStarted","Data":"749f5ecdf364ed49d5420fc8a866b83fcd5c4a5bc7886d25f429d928e70f698f"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.789186 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" event={"ID":"001408b5-e454-432d-93dc-3a2dc497941f","Type":"ContainerStarted","Data":"83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.789240 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" event={"ID":"001408b5-e454-432d-93dc-3a2dc497941f","Type":"ContainerStarted","Data":"4c55978594c509d2aa4b298fdc59a5cdd9768cc904a52ab9627ee016674e5012"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.789970 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.794283 4987 generic.go:334] "Generic (PLEG): container finished" podID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerID="cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb" exitCode=0 Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.794347 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m92l5" event={"ID":"b0640587-71f4-44e4-a6a5-5dc267d57844","Type":"ContainerDied","Data":"cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.794371 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m92l5" event={"ID":"b0640587-71f4-44e4-a6a5-5dc267d57844","Type":"ContainerStarted","Data":"b8e83270e1b36ec7e88a4540833f90538ad3ebd394e9f88406d16520aae8cfe3"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.817362 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.824210 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:20 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:20 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:20 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.824300 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.827042 4987 generic.go:334] "Generic (PLEG): container finished" podID="3d5f0ecd-6403-4646-9197-3826ef18b8f6" containerID="058e4c1b620d266b290dc5386907af14280e947296843604ad440df04a8235a9" exitCode=0 Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.828850 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d5f0ecd-6403-4646-9197-3826ef18b8f6","Type":"ContainerDied","Data":"058e4c1b620d266b290dc5386907af14280e947296843604ad440df04a8235a9"} Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.835815 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wxvx4" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.836558 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" podStartSLOduration=136.836542519 podStartE2EDuration="2m16.836542519s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:20.819556742 +0000 UTC m=+157.874455584" watchObservedRunningTime="2025-10-04 07:12:20.836542519 +0000 UTC m=+157.891441361" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.872205 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-nv4lt" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.876362 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7sbxk" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.883958 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rzqt7"] Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.886323 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.889831 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.895553 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzqt7"] Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.955374 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:12:20 crc kubenswrapper[4987]: I1004 07:12:20.973215 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vw86c" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.055276 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-utilities\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.055358 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-catalog-content\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.055377 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfhq5\" (UniqueName: \"kubernetes.io/projected/6bfd336f-9602-43a5-b18e-2f7d24005d5a-kube-api-access-zfhq5\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.157174 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-utilities\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.157238 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-catalog-content\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.157267 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfhq5\" (UniqueName: \"kubernetes.io/projected/6bfd336f-9602-43a5-b18e-2f7d24005d5a-kube-api-access-zfhq5\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.157923 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-utilities\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.158673 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-catalog-content\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.183324 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfhq5\" (UniqueName: \"kubernetes.io/projected/6bfd336f-9602-43a5-b18e-2f7d24005d5a-kube-api-access-zfhq5\") pod \"redhat-operators-rzqt7\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.262733 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w2qfp"] Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.264417 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.265926 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.277505 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2qfp"] Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.359935 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-utilities\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.359984 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-catalog-content\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.360036 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbqcj\" (UniqueName: \"kubernetes.io/projected/85b17e07-a883-44ac-8f1c-2c98b6eff18f-kube-api-access-bbqcj\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.461502 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-utilities\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.461551 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-catalog-content\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.461598 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbqcj\" (UniqueName: \"kubernetes.io/projected/85b17e07-a883-44ac-8f1c-2c98b6eff18f-kube-api-access-bbqcj\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.463669 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-catalog-content\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.464222 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-utilities\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.482336 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbqcj\" (UniqueName: \"kubernetes.io/projected/85b17e07-a883-44ac-8f1c-2c98b6eff18f-kube-api-access-bbqcj\") pod \"redhat-operators-w2qfp\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.569124 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rzqt7"] Oct 04 07:12:21 crc kubenswrapper[4987]: W1004 07:12:21.585005 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bfd336f_9602_43a5_b18e_2f7d24005d5a.slice/crio-3152a3b57c2f77ed8661b13ed225f8f6292b2a06b1ab2a25385f9088e60d9d52 WatchSource:0}: Error finding container 3152a3b57c2f77ed8661b13ed225f8f6292b2a06b1ab2a25385f9088e60d9d52: Status 404 returned error can't find the container with id 3152a3b57c2f77ed8661b13ed225f8f6292b2a06b1ab2a25385f9088e60d9d52 Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.597551 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.663479 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.664326 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.666666 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.667413 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.689662 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.764136 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4966067c-6290-47b9-bf43-bbb1cc94b24d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.764601 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4966067c-6290-47b9-bf43-bbb1cc94b24d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.821379 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:21 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:21 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:21 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.821459 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.860616 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzqt7" event={"ID":"6bfd336f-9602-43a5-b18e-2f7d24005d5a","Type":"ContainerStarted","Data":"3152a3b57c2f77ed8661b13ed225f8f6292b2a06b1ab2a25385f9088e60d9d52"} Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.866038 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4966067c-6290-47b9-bf43-bbb1cc94b24d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.866091 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4966067c-6290-47b9-bf43-bbb1cc94b24d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.866227 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4966067c-6290-47b9-bf43-bbb1cc94b24d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.886617 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4966067c-6290-47b9-bf43-bbb1cc94b24d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:21 crc kubenswrapper[4987]: I1004 07:12:21.980930 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w2qfp"] Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.056713 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.262791 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.393945 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kube-api-access\") pod \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.394106 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kubelet-dir\") pod \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\" (UID: \"3d5f0ecd-6403-4646-9197-3826ef18b8f6\") " Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.394422 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3d5f0ecd-6403-4646-9197-3826ef18b8f6" (UID: "3d5f0ecd-6403-4646-9197-3826ef18b8f6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.401282 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3d5f0ecd-6403-4646-9197-3826ef18b8f6" (UID: "3d5f0ecd-6403-4646-9197-3826ef18b8f6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.497605 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.497652 4987 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d5f0ecd-6403-4646-9197-3826ef18b8f6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.604060 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 07:12:22 crc kubenswrapper[4987]: W1004 07:12:22.650169 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4966067c_6290_47b9_bf43_bbb1cc94b24d.slice/crio-98d149f05334a38818be222e8f22620d40a747a87eab44c3b4288fc970fd01ab WatchSource:0}: Error finding container 98d149f05334a38818be222e8f22620d40a747a87eab44c3b4288fc970fd01ab: Status 404 returned error can't find the container with id 98d149f05334a38818be222e8f22620d40a747a87eab44c3b4288fc970fd01ab Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.708236 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-nv9l8" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.823038 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:22 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:22 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:22 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.823132 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.885970 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3d5f0ecd-6403-4646-9197-3826ef18b8f6","Type":"ContainerDied","Data":"58d03bd3b00c52f9577a5c5ddcac5e9e0499b07cc6edc84c46161626b38a60ab"} Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.886113 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58d03bd3b00c52f9577a5c5ddcac5e9e0499b07cc6edc84c46161626b38a60ab" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.886026 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.888303 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4966067c-6290-47b9-bf43-bbb1cc94b24d","Type":"ContainerStarted","Data":"98d149f05334a38818be222e8f22620d40a747a87eab44c3b4288fc970fd01ab"} Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.891469 4987 generic.go:334] "Generic (PLEG): container finished" podID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerID="519342791b021bbc00fe04ddc0784cd24fa3f662b598d017423d3367d01703ee" exitCode=0 Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.891572 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzqt7" event={"ID":"6bfd336f-9602-43a5-b18e-2f7d24005d5a","Type":"ContainerDied","Data":"519342791b021bbc00fe04ddc0784cd24fa3f662b598d017423d3367d01703ee"} Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.914798 4987 generic.go:334] "Generic (PLEG): container finished" podID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerID="121e319bff0d0c57f113eaeb7c0f2feb4919eb236d17a16a5a5528baec3794e3" exitCode=0 Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.914923 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerDied","Data":"121e319bff0d0c57f113eaeb7c0f2feb4919eb236d17a16a5a5528baec3794e3"} Oct 04 07:12:22 crc kubenswrapper[4987]: I1004 07:12:22.915235 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerStarted","Data":"6c94b690b8bd42e93e1f78ecd71a59db0d259ca56d94a0c8bcb21474c2ebbf82"} Oct 04 07:12:23 crc kubenswrapper[4987]: I1004 07:12:23.823061 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:23 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:23 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:23 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:23 crc kubenswrapper[4987]: I1004 07:12:23.823509 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:23 crc kubenswrapper[4987]: I1004 07:12:23.939263 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4966067c-6290-47b9-bf43-bbb1cc94b24d","Type":"ContainerStarted","Data":"82b34ef42daf0b34147353c2d4157c50c5bfd16711cad3d4315a8728c2273aeb"} Oct 04 07:12:23 crc kubenswrapper[4987]: I1004 07:12:23.960771 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.96074898 podStartE2EDuration="2.96074898s" podCreationTimestamp="2025-10-04 07:12:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:23.954759078 +0000 UTC m=+161.009657920" watchObservedRunningTime="2025-10-04 07:12:23.96074898 +0000 UTC m=+161.015647822" Oct 04 07:12:24 crc kubenswrapper[4987]: I1004 07:12:24.819847 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:24 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:24 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:24 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:24 crc kubenswrapper[4987]: I1004 07:12:24.819916 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:24 crc kubenswrapper[4987]: I1004 07:12:24.917880 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:24 crc kubenswrapper[4987]: I1004 07:12:24.931912 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-wt287" Oct 04 07:12:25 crc kubenswrapper[4987]: I1004 07:12:25.084770 4987 generic.go:334] "Generic (PLEG): container finished" podID="4966067c-6290-47b9-bf43-bbb1cc94b24d" containerID="82b34ef42daf0b34147353c2d4157c50c5bfd16711cad3d4315a8728c2273aeb" exitCode=0 Oct 04 07:12:25 crc kubenswrapper[4987]: I1004 07:12:25.085847 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4966067c-6290-47b9-bf43-bbb1cc94b24d","Type":"ContainerDied","Data":"82b34ef42daf0b34147353c2d4157c50c5bfd16711cad3d4315a8728c2273aeb"} Oct 04 07:12:25 crc kubenswrapper[4987]: I1004 07:12:25.821977 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:25 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:25 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:25 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:25 crc kubenswrapper[4987]: I1004 07:12:25.822047 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:26 crc kubenswrapper[4987]: I1004 07:12:26.822313 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:26 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:26 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:26 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:26 crc kubenswrapper[4987]: I1004 07:12:26.822785 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:27 crc kubenswrapper[4987]: I1004 07:12:27.820689 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:27 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:27 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:27 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:27 crc kubenswrapper[4987]: I1004 07:12:27.820790 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:28 crc kubenswrapper[4987]: I1004 07:12:28.395827 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:12:28 crc kubenswrapper[4987]: I1004 07:12:28.403697 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4b4d7a8a-0a67-4831-9a78-1369771b81db-metrics-certs\") pod \"network-metrics-daemon-m6jjp\" (UID: \"4b4d7a8a-0a67-4831-9a78-1369771b81db\") " pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:12:28 crc kubenswrapper[4987]: I1004 07:12:28.674695 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m6jjp" Oct 04 07:12:28 crc kubenswrapper[4987]: I1004 07:12:28.820286 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:28 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:28 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:28 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:28 crc kubenswrapper[4987]: I1004 07:12:28.820341 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:29 crc kubenswrapper[4987]: I1004 07:12:29.820449 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:29 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:29 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:29 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:29 crc kubenswrapper[4987]: I1004 07:12:29.820536 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.117197 4987 patch_prober.go:28] interesting pod/downloads-7954f5f757-t2w8v container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.117282 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-t2w8v" podUID="453b9a94-8a64-477b-97eb-633594674f79" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.117322 4987 patch_prober.go:28] interesting pod/downloads-7954f5f757-t2w8v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.117413 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-t2w8v" podUID="453b9a94-8a64-477b-97eb-633594674f79" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.343531 4987 patch_prober.go:28] interesting pod/console-f9d7485db-wfp5w container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.343591 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wfp5w" podUID="dc8c0c59-a185-4b91-920e-c53f58d27e61" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.819754 4987 patch_prober.go:28] interesting pod/router-default-5444994796-bjgbf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 07:12:30 crc kubenswrapper[4987]: [-]has-synced failed: reason withheld Oct 04 07:12:30 crc kubenswrapper[4987]: [+]process-running ok Oct 04 07:12:30 crc kubenswrapper[4987]: healthz check failed Oct 04 07:12:30 crc kubenswrapper[4987]: I1004 07:12:30.819908 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bjgbf" podUID="3e1f11da-af03-4bc8-97c8-ad2eac21d489" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.227813 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.333222 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4966067c-6290-47b9-bf43-bbb1cc94b24d-kube-api-access\") pod \"4966067c-6290-47b9-bf43-bbb1cc94b24d\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.333284 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4966067c-6290-47b9-bf43-bbb1cc94b24d-kubelet-dir\") pod \"4966067c-6290-47b9-bf43-bbb1cc94b24d\" (UID: \"4966067c-6290-47b9-bf43-bbb1cc94b24d\") " Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.333419 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4966067c-6290-47b9-bf43-bbb1cc94b24d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4966067c-6290-47b9-bf43-bbb1cc94b24d" (UID: "4966067c-6290-47b9-bf43-bbb1cc94b24d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.333682 4987 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4966067c-6290-47b9-bf43-bbb1cc94b24d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.353915 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4966067c-6290-47b9-bf43-bbb1cc94b24d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4966067c-6290-47b9-bf43-bbb1cc94b24d" (UID: "4966067c-6290-47b9-bf43-bbb1cc94b24d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.435350 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4966067c-6290-47b9-bf43-bbb1cc94b24d-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.824284 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:31 crc kubenswrapper[4987]: I1004 07:12:31.826596 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bjgbf" Oct 04 07:12:32 crc kubenswrapper[4987]: I1004 07:12:32.162815 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4966067c-6290-47b9-bf43-bbb1cc94b24d","Type":"ContainerDied","Data":"98d149f05334a38818be222e8f22620d40a747a87eab44c3b4288fc970fd01ab"} Oct 04 07:12:32 crc kubenswrapper[4987]: I1004 07:12:32.162884 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98d149f05334a38818be222e8f22620d40a747a87eab44c3b4288fc970fd01ab" Oct 04 07:12:32 crc kubenswrapper[4987]: I1004 07:12:32.163063 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 07:12:39 crc kubenswrapper[4987]: I1004 07:12:39.432332 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:12:40 crc kubenswrapper[4987]: I1004 07:12:40.123276 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-t2w8v" Oct 04 07:12:40 crc kubenswrapper[4987]: I1004 07:12:40.346336 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:40 crc kubenswrapper[4987]: I1004 07:12:40.353346 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wfp5w" Oct 04 07:12:42 crc kubenswrapper[4987]: I1004 07:12:42.703317 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:12:42 crc kubenswrapper[4987]: I1004 07:12:42.703802 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:12:50 crc kubenswrapper[4987]: I1004 07:12:50.903728 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5gz2f" Oct 04 07:12:52 crc kubenswrapper[4987]: E1004 07:12:52.529454 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 07:12:52 crc kubenswrapper[4987]: E1004 07:12:52.529737 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59kfc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6fdjr_openshift-marketplace(bc5dfaf7-a38e-4399-b205-83aeb7adfc8d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:52 crc kubenswrapper[4987]: E1004 07:12:52.531192 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6fdjr" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" Oct 04 07:12:52 crc kubenswrapper[4987]: E1004 07:12:52.567288 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 07:12:52 crc kubenswrapper[4987]: E1004 07:12:52.567527 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lqcbx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-m92l5_openshift-marketplace(b0640587-71f4-44e4-a6a5-5dc267d57844): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:52 crc kubenswrapper[4987]: E1004 07:12:52.568695 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-m92l5" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" Oct 04 07:12:53 crc kubenswrapper[4987]: I1004 07:12:53.098339 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 07:12:53 crc kubenswrapper[4987]: E1004 07:12:53.478488 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6fdjr" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" Oct 04 07:12:53 crc kubenswrapper[4987]: E1004 07:12:53.478818 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-m92l5" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" Oct 04 07:12:53 crc kubenswrapper[4987]: E1004 07:12:53.546857 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 07:12:53 crc kubenswrapper[4987]: E1004 07:12:53.547008 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4x9t2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8lddn_openshift-marketplace(c5530de4-963e-4282-9e46-5ead5330121f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:53 crc kubenswrapper[4987]: E1004 07:12:53.548258 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8lddn" podUID="c5530de4-963e-4282-9e46-5ead5330121f" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.784862 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8lddn" podUID="c5530de4-963e-4282-9e46-5ead5330121f" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.845269 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.845429 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p8zgq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gls68_openshift-marketplace(f0cd1c87-9177-4edd-ab2d-453d8ba8b4af): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.846813 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gls68" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.883753 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.883904 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cwldq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-98fbt_openshift-marketplace(2841f4fc-7a32-4d8a-91a7-f088396830f3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:54 crc kubenswrapper[4987]: E1004 07:12:54.885241 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-98fbt" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.618123 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gls68" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.618189 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-98fbt" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.694287 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.694682 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bbqcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w2qfp_openshift-marketplace(85b17e07-a883-44ac-8f1c-2c98b6eff18f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.695920 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w2qfp" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.739875 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.740054 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vkgqs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-p9k8v_openshift-marketplace(36b7fa18-bb86-4094-b3a8-2e8323fa2a05): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.740288 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.740341 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zfhq5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rzqt7_openshift-marketplace(6bfd336f-9602-43a5-b18e-2f7d24005d5a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.741194 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-p9k8v" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" Oct 04 07:12:57 crc kubenswrapper[4987]: E1004 07:12:57.741582 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-rzqt7" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" Oct 04 07:12:58 crc kubenswrapper[4987]: I1004 07:12:58.061141 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m6jjp"] Oct 04 07:12:58 crc kubenswrapper[4987]: I1004 07:12:58.312559 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" event={"ID":"4b4d7a8a-0a67-4831-9a78-1369771b81db","Type":"ContainerStarted","Data":"f01a4ad48ebad4e582cdb9c96ec255607192fcb30dffd24097d2bae79597ce33"} Oct 04 07:12:58 crc kubenswrapper[4987]: E1004 07:12:58.315340 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w2qfp" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" Oct 04 07:12:58 crc kubenswrapper[4987]: E1004 07:12:58.315407 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-p9k8v" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" Oct 04 07:12:58 crc kubenswrapper[4987]: E1004 07:12:58.316466 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rzqt7" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" Oct 04 07:12:59 crc kubenswrapper[4987]: I1004 07:12:59.323737 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" event={"ID":"4b4d7a8a-0a67-4831-9a78-1369771b81db","Type":"ContainerStarted","Data":"95c30758c589a1e03bc1b77020ee12af0c03a1684b1976bd233e1eb2cddf0f57"} Oct 04 07:12:59 crc kubenswrapper[4987]: I1004 07:12:59.324044 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m6jjp" event={"ID":"4b4d7a8a-0a67-4831-9a78-1369771b81db","Type":"ContainerStarted","Data":"0e698404b5f5e615954520e5943d8f75cb5504da32c689993a41d7d2afb8025b"} Oct 04 07:12:59 crc kubenswrapper[4987]: I1004 07:12:59.348987 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-m6jjp" podStartSLOduration=175.348946014 podStartE2EDuration="2m55.348946014s" podCreationTimestamp="2025-10-04 07:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:12:59.344090147 +0000 UTC m=+196.398988999" watchObservedRunningTime="2025-10-04 07:12:59.348946014 +0000 UTC m=+196.403844866" Oct 04 07:13:06 crc kubenswrapper[4987]: I1004 07:13:06.361042 4987 generic.go:334] "Generic (PLEG): container finished" podID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerID="f8e313c7303e6dabb7d397cb9e9f4809b46034a0f85b55f339438296a5527a07" exitCode=0 Oct 04 07:13:06 crc kubenswrapper[4987]: I1004 07:13:06.361106 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fdjr" event={"ID":"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d","Type":"ContainerDied","Data":"f8e313c7303e6dabb7d397cb9e9f4809b46034a0f85b55f339438296a5527a07"} Oct 04 07:13:07 crc kubenswrapper[4987]: I1004 07:13:07.370663 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fdjr" event={"ID":"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d","Type":"ContainerStarted","Data":"792bc41498c222d9740dca93b8494b6a263702c167aba9adcd11c56c6dfa38f7"} Oct 04 07:13:07 crc kubenswrapper[4987]: I1004 07:13:07.391706 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6fdjr" podStartSLOduration=1.961641285 podStartE2EDuration="48.391679476s" podCreationTimestamp="2025-10-04 07:12:19 +0000 UTC" firstStartedPulling="2025-10-04 07:12:20.779692278 +0000 UTC m=+157.834591120" lastFinishedPulling="2025-10-04 07:13:07.209730469 +0000 UTC m=+204.264629311" observedRunningTime="2025-10-04 07:13:07.384179872 +0000 UTC m=+204.439078714" watchObservedRunningTime="2025-10-04 07:13:07.391679476 +0000 UTC m=+204.446578318" Oct 04 07:13:09 crc kubenswrapper[4987]: I1004 07:13:09.388021 4987 generic.go:334] "Generic (PLEG): container finished" podID="c5530de4-963e-4282-9e46-5ead5330121f" containerID="6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0" exitCode=0 Oct 04 07:13:09 crc kubenswrapper[4987]: I1004 07:13:09.388479 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerDied","Data":"6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0"} Oct 04 07:13:09 crc kubenswrapper[4987]: I1004 07:13:09.395087 4987 generic.go:334] "Generic (PLEG): container finished" podID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerID="42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386" exitCode=0 Oct 04 07:13:09 crc kubenswrapper[4987]: I1004 07:13:09.395138 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m92l5" event={"ID":"b0640587-71f4-44e4-a6a5-5dc267d57844","Type":"ContainerDied","Data":"42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386"} Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.184677 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.185064 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.311111 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.402514 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m92l5" event={"ID":"b0640587-71f4-44e4-a6a5-5dc267d57844","Type":"ContainerStarted","Data":"a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98"} Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.404844 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerStarted","Data":"69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598"} Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.442307 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8lddn" podStartSLOduration=3.369706033 podStartE2EDuration="53.442291987s" podCreationTimestamp="2025-10-04 07:12:17 +0000 UTC" firstStartedPulling="2025-10-04 07:12:19.723741914 +0000 UTC m=+156.778640756" lastFinishedPulling="2025-10-04 07:13:09.796327848 +0000 UTC m=+206.851226710" observedRunningTime="2025-10-04 07:13:10.438971788 +0000 UTC m=+207.493870630" watchObservedRunningTime="2025-10-04 07:13:10.442291987 +0000 UTC m=+207.497190829" Oct 04 07:13:10 crc kubenswrapper[4987]: I1004 07:13:10.443467 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m92l5" podStartSLOduration=2.473072834 podStartE2EDuration="51.443462255s" podCreationTimestamp="2025-10-04 07:12:19 +0000 UTC" firstStartedPulling="2025-10-04 07:12:20.795852024 +0000 UTC m=+157.850750866" lastFinishedPulling="2025-10-04 07:13:09.766241425 +0000 UTC m=+206.821140287" observedRunningTime="2025-10-04 07:13:10.420833585 +0000 UTC m=+207.475732437" watchObservedRunningTime="2025-10-04 07:13:10.443462255 +0000 UTC m=+207.498361097" Oct 04 07:13:11 crc kubenswrapper[4987]: I1004 07:13:11.421189 4987 generic.go:334] "Generic (PLEG): container finished" podID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerID="35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456" exitCode=0 Oct 04 07:13:11 crc kubenswrapper[4987]: I1004 07:13:11.421268 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gls68" event={"ID":"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af","Type":"ContainerDied","Data":"35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456"} Oct 04 07:13:11 crc kubenswrapper[4987]: I1004 07:13:11.423860 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerStarted","Data":"f038a51e699df97bd16c4170d86e78b81b9d4a38597095c47a9bd918514517d0"} Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.430982 4987 generic.go:334] "Generic (PLEG): container finished" podID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerID="4082ed8c55de25da361b9efa12a7570622a05f4804d14cbd931ee99b28f4b712" exitCode=0 Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.431068 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9k8v" event={"ID":"36b7fa18-bb86-4094-b3a8-2e8323fa2a05","Type":"ContainerDied","Data":"4082ed8c55de25da361b9efa12a7570622a05f4804d14cbd931ee99b28f4b712"} Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.441974 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gls68" event={"ID":"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af","Type":"ContainerStarted","Data":"635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638"} Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.454678 4987 generic.go:334] "Generic (PLEG): container finished" podID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerID="f038a51e699df97bd16c4170d86e78b81b9d4a38597095c47a9bd918514517d0" exitCode=0 Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.454757 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerDied","Data":"f038a51e699df97bd16c4170d86e78b81b9d4a38597095c47a9bd918514517d0"} Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.474692 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gls68" podStartSLOduration=2.380447897 podStartE2EDuration="54.474673409s" podCreationTimestamp="2025-10-04 07:12:18 +0000 UTC" firstStartedPulling="2025-10-04 07:12:19.754300937 +0000 UTC m=+156.809199779" lastFinishedPulling="2025-10-04 07:13:11.848526449 +0000 UTC m=+208.903425291" observedRunningTime="2025-10-04 07:13:12.472227379 +0000 UTC m=+209.527126221" watchObservedRunningTime="2025-10-04 07:13:12.474673409 +0000 UTC m=+209.529572251" Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.702890 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.702956 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.703003 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.703711 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:13:12 crc kubenswrapper[4987]: I1004 07:13:12.703840 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910" gracePeriod=600 Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.463323 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerStarted","Data":"f22e2091e882b569ede294f84f0e2c46a8212d9dd5025e8c8989ba3c05ec6b47"} Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.465361 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9k8v" event={"ID":"36b7fa18-bb86-4094-b3a8-2e8323fa2a05","Type":"ContainerStarted","Data":"8cdb97fbd07eeac2d85a47c15f81e97e49884a45da86c3b57ef9c84ae5404669"} Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.467183 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910" exitCode=0 Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.467276 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910"} Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.467358 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"d1f7c32b490273e9ddd0a08c984ac824a51480ca48e9c2de0a6c24f34a688a2e"} Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.469053 4987 generic.go:334] "Generic (PLEG): container finished" podID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerID="1db5992996b1526fea1643701e34958996048978e4464ec763e8188f991ce218" exitCode=0 Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.469100 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbt" event={"ID":"2841f4fc-7a32-4d8a-91a7-f088396830f3","Type":"ContainerDied","Data":"1db5992996b1526fea1643701e34958996048978e4464ec763e8188f991ce218"} Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.518396 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w2qfp" podStartSLOduration=2.52417182 podStartE2EDuration="52.518374849s" podCreationTimestamp="2025-10-04 07:12:21 +0000 UTC" firstStartedPulling="2025-10-04 07:12:22.918893652 +0000 UTC m=+159.973792494" lastFinishedPulling="2025-10-04 07:13:12.913096681 +0000 UTC m=+209.967995523" observedRunningTime="2025-10-04 07:13:13.491057786 +0000 UTC m=+210.545956628" watchObservedRunningTime="2025-10-04 07:13:13.518374849 +0000 UTC m=+210.573273691" Oct 04 07:13:13 crc kubenswrapper[4987]: I1004 07:13:13.536534 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p9k8v" podStartSLOduration=3.409917809 podStartE2EDuration="56.536511361s" podCreationTimestamp="2025-10-04 07:12:17 +0000 UTC" firstStartedPulling="2025-10-04 07:12:19.739149531 +0000 UTC m=+156.794048373" lastFinishedPulling="2025-10-04 07:13:12.865743083 +0000 UTC m=+209.920641925" observedRunningTime="2025-10-04 07:13:13.533803583 +0000 UTC m=+210.588702425" watchObservedRunningTime="2025-10-04 07:13:13.536511361 +0000 UTC m=+210.591410203" Oct 04 07:13:15 crc kubenswrapper[4987]: I1004 07:13:15.483691 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbt" event={"ID":"2841f4fc-7a32-4d8a-91a7-f088396830f3","Type":"ContainerStarted","Data":"7be14b091da2f50792d72dbf8b344ae7ac96817713ffcd372b62c85470bd7bab"} Oct 04 07:13:15 crc kubenswrapper[4987]: I1004 07:13:15.486185 4987 generic.go:334] "Generic (PLEG): container finished" podID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerID="54ea90b05288f109ced387a14e97c6464fe5f0d67d826fd6e99faf46dfc97652" exitCode=0 Oct 04 07:13:15 crc kubenswrapper[4987]: I1004 07:13:15.486247 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzqt7" event={"ID":"6bfd336f-9602-43a5-b18e-2f7d24005d5a","Type":"ContainerDied","Data":"54ea90b05288f109ced387a14e97c6464fe5f0d67d826fd6e99faf46dfc97652"} Oct 04 07:13:15 crc kubenswrapper[4987]: I1004 07:13:15.505887 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-98fbt" podStartSLOduration=3.706651286 podStartE2EDuration="58.505869484s" podCreationTimestamp="2025-10-04 07:12:17 +0000 UTC" firstStartedPulling="2025-10-04 07:12:19.750127106 +0000 UTC m=+156.805025948" lastFinishedPulling="2025-10-04 07:13:14.549345304 +0000 UTC m=+211.604244146" observedRunningTime="2025-10-04 07:13:15.504716916 +0000 UTC m=+212.559615748" watchObservedRunningTime="2025-10-04 07:13:15.505869484 +0000 UTC m=+212.560768326" Oct 04 07:13:16 crc kubenswrapper[4987]: I1004 07:13:16.494471 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzqt7" event={"ID":"6bfd336f-9602-43a5-b18e-2f7d24005d5a","Type":"ContainerStarted","Data":"e7bda8f7a5d29a75c07a5719638b10e2f51ece6f16031eb8853e9df796ca3bac"} Oct 04 07:13:16 crc kubenswrapper[4987]: I1004 07:13:16.514560 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rzqt7" podStartSLOduration=3.441682077 podStartE2EDuration="56.514538799s" podCreationTimestamp="2025-10-04 07:12:20 +0000 UTC" firstStartedPulling="2025-10-04 07:12:22.902869662 +0000 UTC m=+159.957768504" lastFinishedPulling="2025-10-04 07:13:15.975726384 +0000 UTC m=+213.030625226" observedRunningTime="2025-10-04 07:13:16.513896268 +0000 UTC m=+213.568795110" watchObservedRunningTime="2025-10-04 07:13:16.514538799 +0000 UTC m=+213.569437641" Oct 04 07:13:17 crc kubenswrapper[4987]: I1004 07:13:17.825405 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:13:17 crc kubenswrapper[4987]: I1004 07:13:17.826011 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:13:17 crc kubenswrapper[4987]: I1004 07:13:17.880384 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.010680 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.010717 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.059846 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.230175 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.230246 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.274027 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.545236 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.545316 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.701522 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.701579 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:13:18 crc kubenswrapper[4987]: I1004 07:13:18.745884 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:13:19 crc kubenswrapper[4987]: I1004 07:13:19.560092 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:13:19 crc kubenswrapper[4987]: I1004 07:13:19.834943 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:13:19 crc kubenswrapper[4987]: I1004 07:13:19.835357 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:13:19 crc kubenswrapper[4987]: I1004 07:13:19.872011 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:13:20 crc kubenswrapper[4987]: I1004 07:13:20.225158 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:13:20 crc kubenswrapper[4987]: I1004 07:13:20.396291 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gls68"] Oct 04 07:13:20 crc kubenswrapper[4987]: I1004 07:13:20.566906 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:13:20 crc kubenswrapper[4987]: I1004 07:13:20.997651 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9k8v"] Oct 04 07:13:20 crc kubenswrapper[4987]: I1004 07:13:20.997905 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p9k8v" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="registry-server" containerID="cri-o://8cdb97fbd07eeac2d85a47c15f81e97e49884a45da86c3b57ef9c84ae5404669" gracePeriod=2 Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.266930 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.267005 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.314957 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.534565 4987 generic.go:334] "Generic (PLEG): container finished" podID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerID="8cdb97fbd07eeac2d85a47c15f81e97e49884a45da86c3b57ef9c84ae5404669" exitCode=0 Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.534747 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9k8v" event={"ID":"36b7fa18-bb86-4094-b3a8-2e8323fa2a05","Type":"ContainerDied","Data":"8cdb97fbd07eeac2d85a47c15f81e97e49884a45da86c3b57ef9c84ae5404669"} Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.535103 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gls68" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="registry-server" containerID="cri-o://635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638" gracePeriod=2 Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.575535 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.598645 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.598695 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:13:21 crc kubenswrapper[4987]: I1004 07:13:21.706601 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.251250 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.399092 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-catalog-content\") pod \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.399143 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkgqs\" (UniqueName: \"kubernetes.io/projected/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-kube-api-access-vkgqs\") pod \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.399234 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-utilities\") pod \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\" (UID: \"36b7fa18-bb86-4094-b3a8-2e8323fa2a05\") " Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.400131 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-utilities" (OuterVolumeSpecName: "utilities") pod "36b7fa18-bb86-4094-b3a8-2e8323fa2a05" (UID: "36b7fa18-bb86-4094-b3a8-2e8323fa2a05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.404427 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-kube-api-access-vkgqs" (OuterVolumeSpecName: "kube-api-access-vkgqs") pod "36b7fa18-bb86-4094-b3a8-2e8323fa2a05" (UID: "36b7fa18-bb86-4094-b3a8-2e8323fa2a05"). InnerVolumeSpecName "kube-api-access-vkgqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.450106 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36b7fa18-bb86-4094-b3a8-2e8323fa2a05" (UID: "36b7fa18-bb86-4094-b3a8-2e8323fa2a05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.478853 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.500365 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.500659 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.500670 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkgqs\" (UniqueName: \"kubernetes.io/projected/36b7fa18-bb86-4094-b3a8-2e8323fa2a05-kube-api-access-vkgqs\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.542112 4987 generic.go:334] "Generic (PLEG): container finished" podID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerID="635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638" exitCode=0 Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.542240 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gls68" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.542782 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gls68" event={"ID":"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af","Type":"ContainerDied","Data":"635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638"} Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.542813 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gls68" event={"ID":"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af","Type":"ContainerDied","Data":"5e7b9148a9f2831edce4b7a1c287269404ec1783d4dae8c225a3850f7bdb072d"} Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.542830 4987 scope.go:117] "RemoveContainer" containerID="635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.546698 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9k8v" event={"ID":"36b7fa18-bb86-4094-b3a8-2e8323fa2a05","Type":"ContainerDied","Data":"7fbdae595c56270c975ab4b2151e36a7e1658c714323d85e94c90713eef04430"} Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.546928 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9k8v" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.566530 4987 scope.go:117] "RemoveContainer" containerID="35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.591165 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9k8v"] Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.595724 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p9k8v"] Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.596986 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.597886 4987 scope.go:117] "RemoveContainer" containerID="cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.601442 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-utilities\") pod \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.602175 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8zgq\" (UniqueName: \"kubernetes.io/projected/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-kube-api-access-p8zgq\") pod \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.602314 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-catalog-content\") pod \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\" (UID: \"f0cd1c87-9177-4edd-ab2d-453d8ba8b4af\") " Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.602592 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-utilities" (OuterVolumeSpecName: "utilities") pod "f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" (UID: "f0cd1c87-9177-4edd-ab2d-453d8ba8b4af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.606352 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-kube-api-access-p8zgq" (OuterVolumeSpecName: "kube-api-access-p8zgq") pod "f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" (UID: "f0cd1c87-9177-4edd-ab2d-453d8ba8b4af"). InnerVolumeSpecName "kube-api-access-p8zgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.635344 4987 scope.go:117] "RemoveContainer" containerID="635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638" Oct 04 07:13:22 crc kubenswrapper[4987]: E1004 07:13:22.635906 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638\": container with ID starting with 635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638 not found: ID does not exist" containerID="635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.635962 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638"} err="failed to get container status \"635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638\": rpc error: code = NotFound desc = could not find container \"635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638\": container with ID starting with 635bda390af7f0a54bd306205be235fb9e08e4b1ed1e68ac7b4818c0b676b638 not found: ID does not exist" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.635997 4987 scope.go:117] "RemoveContainer" containerID="35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456" Oct 04 07:13:22 crc kubenswrapper[4987]: E1004 07:13:22.636470 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456\": container with ID starting with 35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456 not found: ID does not exist" containerID="35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.636499 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456"} err="failed to get container status \"35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456\": rpc error: code = NotFound desc = could not find container \"35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456\": container with ID starting with 35131011ce8202ba8332731f28ceddfcad1d98bbd23d247719592badbfb5b456 not found: ID does not exist" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.636531 4987 scope.go:117] "RemoveContainer" containerID="cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865" Oct 04 07:13:22 crc kubenswrapper[4987]: E1004 07:13:22.636915 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865\": container with ID starting with cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865 not found: ID does not exist" containerID="cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.636965 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865"} err="failed to get container status \"cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865\": rpc error: code = NotFound desc = could not find container \"cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865\": container with ID starting with cd73027ec4a74f27e3ffefb6a20ff3738ecdf966f90e0d41f51411cff9029865 not found: ID does not exist" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.636979 4987 scope.go:117] "RemoveContainer" containerID="8cdb97fbd07eeac2d85a47c15f81e97e49884a45da86c3b57ef9c84ae5404669" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.661809 4987 scope.go:117] "RemoveContainer" containerID="4082ed8c55de25da361b9efa12a7570622a05f4804d14cbd931ee99b28f4b712" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.663067 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" (UID: "f0cd1c87-9177-4edd-ab2d-453d8ba8b4af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.676703 4987 scope.go:117] "RemoveContainer" containerID="8ba6f84942c6cabd9e1b63cb1a11feadc3a3fda213dd8cb39e381cc888e43a96" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.705176 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8zgq\" (UniqueName: \"kubernetes.io/projected/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-kube-api-access-p8zgq\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.705238 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.705247 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.796230 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fdjr"] Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.796502 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6fdjr" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="registry-server" containerID="cri-o://792bc41498c222d9740dca93b8494b6a263702c167aba9adcd11c56c6dfa38f7" gracePeriod=2 Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.883547 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gls68"] Oct 04 07:13:22 crc kubenswrapper[4987]: I1004 07:13:22.886197 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gls68"] Oct 04 07:13:23 crc kubenswrapper[4987]: I1004 07:13:23.554610 4987 generic.go:334] "Generic (PLEG): container finished" podID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerID="792bc41498c222d9740dca93b8494b6a263702c167aba9adcd11c56c6dfa38f7" exitCode=0 Oct 04 07:13:23 crc kubenswrapper[4987]: I1004 07:13:23.554690 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fdjr" event={"ID":"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d","Type":"ContainerDied","Data":"792bc41498c222d9740dca93b8494b6a263702c167aba9adcd11c56c6dfa38f7"} Oct 04 07:13:23 crc kubenswrapper[4987]: I1004 07:13:23.876717 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.023642 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-catalog-content\") pod \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.023720 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59kfc\" (UniqueName: \"kubernetes.io/projected/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-kube-api-access-59kfc\") pod \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.023783 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-utilities\") pod \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\" (UID: \"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d\") " Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.024589 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-utilities" (OuterVolumeSpecName: "utilities") pod "bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" (UID: "bc5dfaf7-a38e-4399-b205-83aeb7adfc8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.029813 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-kube-api-access-59kfc" (OuterVolumeSpecName: "kube-api-access-59kfc") pod "bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" (UID: "bc5dfaf7-a38e-4399-b205-83aeb7adfc8d"). InnerVolumeSpecName "kube-api-access-59kfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.037301 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" (UID: "bc5dfaf7-a38e-4399-b205-83aeb7adfc8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.059331 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" path="/var/lib/kubelet/pods/36b7fa18-bb86-4094-b3a8-2e8323fa2a05/volumes" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.060094 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" path="/var/lib/kubelet/pods/f0cd1c87-9177-4edd-ab2d-453d8ba8b4af/volumes" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.125476 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.125509 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59kfc\" (UniqueName: \"kubernetes.io/projected/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-kube-api-access-59kfc\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.125521 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.565056 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fdjr" event={"ID":"bc5dfaf7-a38e-4399-b205-83aeb7adfc8d","Type":"ContainerDied","Data":"749f5ecdf364ed49d5420fc8a866b83fcd5c4a5bc7886d25f429d928e70f698f"} Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.565125 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fdjr" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.565127 4987 scope.go:117] "RemoveContainer" containerID="792bc41498c222d9740dca93b8494b6a263702c167aba9adcd11c56c6dfa38f7" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.585804 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fdjr"] Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.589166 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fdjr"] Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.589546 4987 scope.go:117] "RemoveContainer" containerID="f8e313c7303e6dabb7d397cb9e9f4809b46034a0f85b55f339438296a5527a07" Oct 04 07:13:24 crc kubenswrapper[4987]: I1004 07:13:24.605289 4987 scope.go:117] "RemoveContainer" containerID="7898568fbe78517afc4139ca937320d296e058a896b7293f5b3e5e73f6b75ca0" Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.397151 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w2qfp"] Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.397680 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w2qfp" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="registry-server" containerID="cri-o://f22e2091e882b569ede294f84f0e2c46a8212d9dd5025e8c8989ba3c05ec6b47" gracePeriod=2 Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.573452 4987 generic.go:334] "Generic (PLEG): container finished" podID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerID="f22e2091e882b569ede294f84f0e2c46a8212d9dd5025e8c8989ba3c05ec6b47" exitCode=0 Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.573518 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerDied","Data":"f22e2091e882b569ede294f84f0e2c46a8212d9dd5025e8c8989ba3c05ec6b47"} Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.800756 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.953767 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-utilities\") pod \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.953851 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-catalog-content\") pod \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.953885 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbqcj\" (UniqueName: \"kubernetes.io/projected/85b17e07-a883-44ac-8f1c-2c98b6eff18f-kube-api-access-bbqcj\") pod \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\" (UID: \"85b17e07-a883-44ac-8f1c-2c98b6eff18f\") " Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.954684 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-utilities" (OuterVolumeSpecName: "utilities") pod "85b17e07-a883-44ac-8f1c-2c98b6eff18f" (UID: "85b17e07-a883-44ac-8f1c-2c98b6eff18f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:25 crc kubenswrapper[4987]: I1004 07:13:25.960783 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85b17e07-a883-44ac-8f1c-2c98b6eff18f-kube-api-access-bbqcj" (OuterVolumeSpecName: "kube-api-access-bbqcj") pod "85b17e07-a883-44ac-8f1c-2c98b6eff18f" (UID: "85b17e07-a883-44ac-8f1c-2c98b6eff18f"). InnerVolumeSpecName "kube-api-access-bbqcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.039327 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85b17e07-a883-44ac-8f1c-2c98b6eff18f" (UID: "85b17e07-a883-44ac-8f1c-2c98b6eff18f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.055205 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.055248 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbqcj\" (UniqueName: \"kubernetes.io/projected/85b17e07-a883-44ac-8f1c-2c98b6eff18f-kube-api-access-bbqcj\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.055262 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85b17e07-a883-44ac-8f1c-2c98b6eff18f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.061151 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" path="/var/lib/kubelet/pods/bc5dfaf7-a38e-4399-b205-83aeb7adfc8d/volumes" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.580711 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w2qfp" event={"ID":"85b17e07-a883-44ac-8f1c-2c98b6eff18f","Type":"ContainerDied","Data":"6c94b690b8bd42e93e1f78ecd71a59db0d259ca56d94a0c8bcb21474c2ebbf82"} Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.580766 4987 scope.go:117] "RemoveContainer" containerID="f22e2091e882b569ede294f84f0e2c46a8212d9dd5025e8c8989ba3c05ec6b47" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.580762 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w2qfp" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.596223 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w2qfp"] Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.596717 4987 scope.go:117] "RemoveContainer" containerID="f038a51e699df97bd16c4170d86e78b81b9d4a38597095c47a9bd918514517d0" Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.599210 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w2qfp"] Oct 04 07:13:26 crc kubenswrapper[4987]: I1004 07:13:26.612079 4987 scope.go:117] "RemoveContainer" containerID="121e319bff0d0c57f113eaeb7c0f2feb4919eb236d17a16a5a5528baec3794e3" Oct 04 07:13:28 crc kubenswrapper[4987]: I1004 07:13:28.059089 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" path="/var/lib/kubelet/pods/85b17e07-a883-44ac-8f1c-2c98b6eff18f/volumes" Oct 04 07:13:28 crc kubenswrapper[4987]: I1004 07:13:28.059686 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:13:29 crc kubenswrapper[4987]: I1004 07:13:29.242224 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6hddn"] Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.287071 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerName="oauth-openshift" containerID="cri-o://eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a" gracePeriod=15 Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.654528 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.692719 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-69fb88d4f9-b64bl"] Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.692982 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.692996 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693013 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693021 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693031 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d5f0ecd-6403-4646-9197-3826ef18b8f6" containerName="pruner" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693039 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d5f0ecd-6403-4646-9197-3826ef18b8f6" containerName="pruner" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693047 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693056 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693072 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693078 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693088 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerName="oauth-openshift" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693093 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerName="oauth-openshift" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693100 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693106 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693113 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4966067c-6290-47b9-bf43-bbb1cc94b24d" containerName="pruner" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693118 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="4966067c-6290-47b9-bf43-bbb1cc94b24d" containerName="pruner" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693126 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693133 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693140 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693146 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693155 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693161 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693167 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693172 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693182 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693187 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="extract-content" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693196 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693202 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="extract-utilities" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.693211 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693217 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693311 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b7fa18-bb86-4094-b3a8-2e8323fa2a05" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693322 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0cd1c87-9177-4edd-ab2d-453d8ba8b4af" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693333 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="4966067c-6290-47b9-bf43-bbb1cc94b24d" containerName="pruner" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693341 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="85b17e07-a883-44ac-8f1c-2c98b6eff18f" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693349 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerName="oauth-openshift" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693356 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc5dfaf7-a38e-4399-b205-83aeb7adfc8d" containerName="registry-server" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693365 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d5f0ecd-6403-4646-9197-3826ef18b8f6" containerName="pruner" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.693738 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.694470 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69fb88d4f9-b64bl"] Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.743932 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-serving-cert\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.743996 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-provider-selection\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744017 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-dir\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744034 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-idp-0-file-data\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744054 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-router-certs\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744076 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-cliconfig\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744092 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-trusted-ca-bundle\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744106 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744131 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-service-ca\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744181 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddmnb\" (UniqueName: \"kubernetes.io/projected/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-kube-api-access-ddmnb\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744218 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-ocp-branding-template\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744260 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-policies\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744283 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-session\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744307 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-error\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744356 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-login\") pod \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\" (UID: \"659b7013-28e6-4ccb-b2d6-247c1eff7d4f\") " Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.744590 4987 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.745642 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.745855 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.745928 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.746312 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.749244 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.749258 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.749552 4987 generic.go:334] "Generic (PLEG): container finished" podID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" containerID="eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a" exitCode=0 Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.749611 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" event={"ID":"659b7013-28e6-4ccb-b2d6-247c1eff7d4f","Type":"ContainerDied","Data":"eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a"} Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.749685 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6hddn" event={"ID":"659b7013-28e6-4ccb-b2d6-247c1eff7d4f","Type":"ContainerDied","Data":"5bd6250477f0e9ac51a75b641052bc9bfd3553a9dd86bd3d893d500f00822ecf"} Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.749703 4987 scope.go:117] "RemoveContainer" containerID="eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.750266 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.750516 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-kube-api-access-ddmnb" (OuterVolumeSpecName: "kube-api-access-ddmnb") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "kube-api-access-ddmnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.750886 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.751085 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.751341 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.751571 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.758966 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.759194 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "659b7013-28e6-4ccb-b2d6-247c1eff7d4f" (UID: "659b7013-28e6-4ccb-b2d6-247c1eff7d4f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.790977 4987 scope.go:117] "RemoveContainer" containerID="eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a" Oct 04 07:13:54 crc kubenswrapper[4987]: E1004 07:13:54.791518 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a\": container with ID starting with eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a not found: ID does not exist" containerID="eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.791569 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a"} err="failed to get container status \"eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a\": rpc error: code = NotFound desc = could not find container \"eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a\": container with ID starting with eb3d2fdbb081897433fe347cdf975ac94a3a23b9a2e4de34283aaa8f335e4e6a not found: ID does not exist" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846098 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6gfv\" (UniqueName: \"kubernetes.io/projected/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-kube-api-access-p6gfv\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846375 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-session\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846512 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-audit-policies\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846590 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-service-ca\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846777 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-login\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846859 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-error\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846904 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-audit-dir\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846938 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.846974 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847012 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847051 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847080 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-router-certs\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847154 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847391 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847485 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847503 4987 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847520 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847535 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847547 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847560 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847572 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847585 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847597 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847612 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847640 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847655 4987 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.847668 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddmnb\" (UniqueName: \"kubernetes.io/projected/659b7013-28e6-4ccb-b2d6-247c1eff7d4f-kube-api-access-ddmnb\") on node \"crc\" DevicePath \"\"" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.949252 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-audit-policies\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.949550 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-service-ca\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.949793 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-login\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.949961 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-error\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950080 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-audit-dir\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950195 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950313 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950488 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950668 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950811 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-router-certs\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950964 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.951113 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.951254 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-session\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.951396 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6gfv\" (UniqueName: \"kubernetes.io/projected/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-kube-api-access-p6gfv\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.951740 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950807 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-audit-policies\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.951332 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-audit-dir\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.950747 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-service-ca\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.953226 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-error\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.953985 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-login\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.954286 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-router-certs\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.954666 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.955838 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.956449 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.958360 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.959557 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-session\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.959781 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:54 crc kubenswrapper[4987]: I1004 07:13:54.973585 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6gfv\" (UniqueName: \"kubernetes.io/projected/11ffdd98-a02d-43b7-af6e-4fc37dcc9368-kube-api-access-p6gfv\") pod \"oauth-openshift-69fb88d4f9-b64bl\" (UID: \"11ffdd98-a02d-43b7-af6e-4fc37dcc9368\") " pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.014322 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.090138 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6hddn"] Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.094253 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6hddn"] Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.437181 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69fb88d4f9-b64bl"] Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.757154 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" event={"ID":"11ffdd98-a02d-43b7-af6e-4fc37dcc9368","Type":"ContainerStarted","Data":"bcd593dbc3da645ef40a696b65275c6265432b72838f4ffe110663ee7023b0cd"} Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.757199 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" event={"ID":"11ffdd98-a02d-43b7-af6e-4fc37dcc9368","Type":"ContainerStarted","Data":"9b6e096fe0f28677b340b391611f1e9a6c2a6c179969f10e025d2f35a0f5119a"} Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.757506 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.777949 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" podStartSLOduration=26.777929848 podStartE2EDuration="26.777929848s" podCreationTimestamp="2025-10-04 07:13:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:13:55.772060336 +0000 UTC m=+252.826959208" watchObservedRunningTime="2025-10-04 07:13:55.777929848 +0000 UTC m=+252.832828690" Oct 04 07:13:55 crc kubenswrapper[4987]: I1004 07:13:55.968644 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-69fb88d4f9-b64bl" Oct 04 07:13:56 crc kubenswrapper[4987]: I1004 07:13:56.061496 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="659b7013-28e6-4ccb-b2d6-247c1eff7d4f" path="/var/lib/kubelet/pods/659b7013-28e6-4ccb-b2d6-247c1eff7d4f/volumes" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.325336 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8lddn"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.326693 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8lddn" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="registry-server" containerID="cri-o://69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598" gracePeriod=30 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.338001 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-98fbt"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.338879 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-98fbt" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="registry-server" containerID="cri-o://7be14b091da2f50792d72dbf8b344ae7ac96817713ffcd372b62c85470bd7bab" gracePeriod=30 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.342783 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9wm6"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.343096 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerName="marketplace-operator" containerID="cri-o://b7acd789d89c4ef81a64ca88670dbbbce5fa6aaca5b5c3b741b29cfb477b1c5b" gracePeriod=30 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.349079 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m92l5"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.349454 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m92l5" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="registry-server" containerID="cri-o://a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98" gracePeriod=30 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.366049 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rzqt7"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.366422 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rzqt7" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="registry-server" containerID="cri-o://e7bda8f7a5d29a75c07a5719638b10e2f51ece6f16031eb8853e9df796ca3bac" gracePeriod=30 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.380920 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r46zk"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.381587 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r46zk"] Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.381775 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.553732 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.553800 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.553844 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-666rb\" (UniqueName: \"kubernetes.io/projected/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-kube-api-access-666rb\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.654853 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.655155 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.655199 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-666rb\" (UniqueName: \"kubernetes.io/projected/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-kube-api-access-666rb\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.656063 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.663592 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.671927 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-666rb\" (UniqueName: \"kubernetes.io/projected/4d6cbfa1-a2bb-4c11-87b2-a5897573ed88-kube-api-access-666rb\") pod \"marketplace-operator-79b997595-r46zk\" (UID: \"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88\") " pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.705938 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.846010 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.853240 4987 generic.go:334] "Generic (PLEG): container finished" podID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerID="7be14b091da2f50792d72dbf8b344ae7ac96817713ffcd372b62c85470bd7bab" exitCode=0 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.853397 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbt" event={"ID":"2841f4fc-7a32-4d8a-91a7-f088396830f3","Type":"ContainerDied","Data":"7be14b091da2f50792d72dbf8b344ae7ac96817713ffcd372b62c85470bd7bab"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.853448 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-98fbt" event={"ID":"2841f4fc-7a32-4d8a-91a7-f088396830f3","Type":"ContainerDied","Data":"22b2b5f60d4254cb1b4fb720bbfc834b5766b47f54b400656bdb82664874dac4"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.853468 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22b2b5f60d4254cb1b4fb720bbfc834b5766b47f54b400656bdb82664874dac4" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.859078 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.859190 4987 generic.go:334] "Generic (PLEG): container finished" podID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerID="e7bda8f7a5d29a75c07a5719638b10e2f51ece6f16031eb8853e9df796ca3bac" exitCode=0 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.859258 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzqt7" event={"ID":"6bfd336f-9602-43a5-b18e-2f7d24005d5a","Type":"ContainerDied","Data":"e7bda8f7a5d29a75c07a5719638b10e2f51ece6f16031eb8853e9df796ca3bac"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.859288 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rzqt7" event={"ID":"6bfd336f-9602-43a5-b18e-2f7d24005d5a","Type":"ContainerDied","Data":"3152a3b57c2f77ed8661b13ed225f8f6292b2a06b1ab2a25385f9088e60d9d52"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.859301 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3152a3b57c2f77ed8661b13ed225f8f6292b2a06b1ab2a25385f9088e60d9d52" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.862427 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.862944 4987 generic.go:334] "Generic (PLEG): container finished" podID="c5530de4-963e-4282-9e46-5ead5330121f" containerID="69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598" exitCode=0 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.863000 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerDied","Data":"69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.863029 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lddn" event={"ID":"c5530de4-963e-4282-9e46-5ead5330121f","Type":"ContainerDied","Data":"6235be29ed2a2248e7dba967f6f27c98a26ac3aed22a091b4326e4e1a7d62201"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.863046 4987 scope.go:117] "RemoveContainer" containerID="69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.863178 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lddn" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.880355 4987 generic.go:334] "Generic (PLEG): container finished" podID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerID="a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98" exitCode=0 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.880458 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m92l5" event={"ID":"b0640587-71f4-44e4-a6a5-5dc267d57844","Type":"ContainerDied","Data":"a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.880500 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m92l5" event={"ID":"b0640587-71f4-44e4-a6a5-5dc267d57844","Type":"ContainerDied","Data":"b8e83270e1b36ec7e88a4540833f90538ad3ebd394e9f88406d16520aae8cfe3"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.880608 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m92l5" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.885141 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.886381 4987 generic.go:334] "Generic (PLEG): container finished" podID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerID="b7acd789d89c4ef81a64ca88670dbbbce5fa6aaca5b5c3b741b29cfb477b1c5b" exitCode=0 Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.886464 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" event={"ID":"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a","Type":"ContainerDied","Data":"b7acd789d89c4ef81a64ca88670dbbbce5fa6aaca5b5c3b741b29cfb477b1c5b"} Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.890280 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.919202 4987 scope.go:117] "RemoveContainer" containerID="6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.944168 4987 scope.go:117] "RemoveContainer" containerID="d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.967066 4987 scope.go:117] "RemoveContainer" containerID="69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.967749 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-utilities\") pod \"b0640587-71f4-44e4-a6a5-5dc267d57844\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " Oct 04 07:14:07 crc kubenswrapper[4987]: E1004 07:14:07.967825 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598\": container with ID starting with 69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598 not found: ID does not exist" containerID="69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.967866 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-catalog-content\") pod \"c5530de4-963e-4282-9e46-5ead5330121f\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.967907 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598"} err="failed to get container status \"69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598\": rpc error: code = NotFound desc = could not find container \"69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598\": container with ID starting with 69cde7ca9a2ead70609a6fd2f0dab5a917a611baaa8ae4162f89bade8d965598 not found: ID does not exist" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.967960 4987 scope.go:117] "RemoveContainer" containerID="6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0" Oct 04 07:14:07 crc kubenswrapper[4987]: E1004 07:14:07.968294 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0\": container with ID starting with 6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0 not found: ID does not exist" containerID="6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.968328 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0"} err="failed to get container status \"6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0\": rpc error: code = NotFound desc = could not find container \"6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0\": container with ID starting with 6c4acceeec20e2c025713e1b88bdf782351bfd7ff96c1798a019e3a3282a24b0 not found: ID does not exist" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.968354 4987 scope.go:117] "RemoveContainer" containerID="d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e" Oct 04 07:14:07 crc kubenswrapper[4987]: E1004 07:14:07.968530 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e\": container with ID starting with d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e not found: ID does not exist" containerID="d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.968553 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e"} err="failed to get container status \"d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e\": rpc error: code = NotFound desc = could not find container \"d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e\": container with ID starting with d0b4d2ffd0bd726cb2a61ae51bb32504259c1b1ad6c73d267846483093de6c5e not found: ID does not exist" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.968569 4987 scope.go:117] "RemoveContainer" containerID="a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.968783 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-utilities" (OuterVolumeSpecName: "utilities") pod "b0640587-71f4-44e4-a6a5-5dc267d57844" (UID: "b0640587-71f4-44e4-a6a5-5dc267d57844"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.975760 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqcbx\" (UniqueName: \"kubernetes.io/projected/b0640587-71f4-44e4-a6a5-5dc267d57844-kube-api-access-lqcbx\") pod \"b0640587-71f4-44e4-a6a5-5dc267d57844\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.975820 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-utilities\") pod \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.975896 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfhq5\" (UniqueName: \"kubernetes.io/projected/6bfd336f-9602-43a5-b18e-2f7d24005d5a-kube-api-access-zfhq5\") pod \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.975931 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x9t2\" (UniqueName: \"kubernetes.io/projected/c5530de4-963e-4282-9e46-5ead5330121f-kube-api-access-4x9t2\") pod \"c5530de4-963e-4282-9e46-5ead5330121f\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.975976 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-catalog-content\") pod \"b0640587-71f4-44e4-a6a5-5dc267d57844\" (UID: \"b0640587-71f4-44e4-a6a5-5dc267d57844\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976020 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwldq\" (UniqueName: \"kubernetes.io/projected/2841f4fc-7a32-4d8a-91a7-f088396830f3-kube-api-access-cwldq\") pod \"2841f4fc-7a32-4d8a-91a7-f088396830f3\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976058 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-utilities\") pod \"2841f4fc-7a32-4d8a-91a7-f088396830f3\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976101 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-operator-metrics\") pod \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976137 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-trusted-ca\") pod \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976204 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-catalog-content\") pod \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\" (UID: \"6bfd336f-9602-43a5-b18e-2f7d24005d5a\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976237 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-utilities\") pod \"c5530de4-963e-4282-9e46-5ead5330121f\" (UID: \"c5530de4-963e-4282-9e46-5ead5330121f\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976286 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-catalog-content\") pod \"2841f4fc-7a32-4d8a-91a7-f088396830f3\" (UID: \"2841f4fc-7a32-4d8a-91a7-f088396830f3\") " Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976485 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-utilities" (OuterVolumeSpecName: "utilities") pod "6bfd336f-9602-43a5-b18e-2f7d24005d5a" (UID: "6bfd336f-9602-43a5-b18e-2f7d24005d5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976869 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.976899 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.977553 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-utilities" (OuterVolumeSpecName: "utilities") pod "c5530de4-963e-4282-9e46-5ead5330121f" (UID: "c5530de4-963e-4282-9e46-5ead5330121f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.977574 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-utilities" (OuterVolumeSpecName: "utilities") pod "2841f4fc-7a32-4d8a-91a7-f088396830f3" (UID: "2841f4fc-7a32-4d8a-91a7-f088396830f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.979230 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" (UID: "57ac644c-0670-49bb-b7ce-3d9ecb72fa5a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.981764 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2841f4fc-7a32-4d8a-91a7-f088396830f3-kube-api-access-cwldq" (OuterVolumeSpecName: "kube-api-access-cwldq") pod "2841f4fc-7a32-4d8a-91a7-f088396830f3" (UID: "2841f4fc-7a32-4d8a-91a7-f088396830f3"). InnerVolumeSpecName "kube-api-access-cwldq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.983041 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" (UID: "57ac644c-0670-49bb-b7ce-3d9ecb72fa5a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.983089 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0640587-71f4-44e4-a6a5-5dc267d57844-kube-api-access-lqcbx" (OuterVolumeSpecName: "kube-api-access-lqcbx") pod "b0640587-71f4-44e4-a6a5-5dc267d57844" (UID: "b0640587-71f4-44e4-a6a5-5dc267d57844"). InnerVolumeSpecName "kube-api-access-lqcbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.985640 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bfd336f-9602-43a5-b18e-2f7d24005d5a-kube-api-access-zfhq5" (OuterVolumeSpecName: "kube-api-access-zfhq5") pod "6bfd336f-9602-43a5-b18e-2f7d24005d5a" (UID: "6bfd336f-9602-43a5-b18e-2f7d24005d5a"). InnerVolumeSpecName "kube-api-access-zfhq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.985481 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5530de4-963e-4282-9e46-5ead5330121f-kube-api-access-4x9t2" (OuterVolumeSpecName: "kube-api-access-4x9t2") pod "c5530de4-963e-4282-9e46-5ead5330121f" (UID: "c5530de4-963e-4282-9e46-5ead5330121f"). InnerVolumeSpecName "kube-api-access-4x9t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.988216 4987 scope.go:117] "RemoveContainer" containerID="42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386" Oct 04 07:14:07 crc kubenswrapper[4987]: I1004 07:14:07.993678 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0640587-71f4-44e4-a6a5-5dc267d57844" (UID: "b0640587-71f4-44e4-a6a5-5dc267d57844"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.005804 4987 scope.go:117] "RemoveContainer" containerID="cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.020835 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5530de4-963e-4282-9e46-5ead5330121f" (UID: "c5530de4-963e-4282-9e46-5ead5330121f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.022028 4987 scope.go:117] "RemoveContainer" containerID="a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98" Oct 04 07:14:08 crc kubenswrapper[4987]: E1004 07:14:08.022480 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98\": container with ID starting with a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98 not found: ID does not exist" containerID="a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.022530 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98"} err="failed to get container status \"a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98\": rpc error: code = NotFound desc = could not find container \"a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98\": container with ID starting with a001bf159d8e638ae97f5fababc67f4a09c0dae1a3b84be7e0f32d7bff42ff98 not found: ID does not exist" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.022562 4987 scope.go:117] "RemoveContainer" containerID="42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386" Oct 04 07:14:08 crc kubenswrapper[4987]: E1004 07:14:08.022992 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386\": container with ID starting with 42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386 not found: ID does not exist" containerID="42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.023027 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386"} err="failed to get container status \"42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386\": rpc error: code = NotFound desc = could not find container \"42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386\": container with ID starting with 42097e2274c71e28d81df27d999c8dc95cac92a027cde594fde5859ea3f44386 not found: ID does not exist" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.023050 4987 scope.go:117] "RemoveContainer" containerID="cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb" Oct 04 07:14:08 crc kubenswrapper[4987]: E1004 07:14:08.023287 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb\": container with ID starting with cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb not found: ID does not exist" containerID="cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.023315 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb"} err="failed to get container status \"cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb\": rpc error: code = NotFound desc = could not find container \"cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb\": container with ID starting with cad56c1ca54df760bb538dfd99c53e5b074e828a6a583808d356b9106af0d3cb not found: ID does not exist" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.042679 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2841f4fc-7a32-4d8a-91a7-f088396830f3" (UID: "2841f4fc-7a32-4d8a-91a7-f088396830f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.059729 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6bfd336f-9602-43a5-b18e-2f7d24005d5a" (UID: "6bfd336f-9602-43a5-b18e-2f7d24005d5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077395 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgbz9\" (UniqueName: \"kubernetes.io/projected/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-kube-api-access-mgbz9\") pod \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\" (UID: \"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a\") " Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077713 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0640587-71f4-44e4-a6a5-5dc267d57844-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077741 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwldq\" (UniqueName: \"kubernetes.io/projected/2841f4fc-7a32-4d8a-91a7-f088396830f3-kube-api-access-cwldq\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077801 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077815 4987 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077825 4987 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077836 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bfd336f-9602-43a5-b18e-2f7d24005d5a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077849 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077859 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2841f4fc-7a32-4d8a-91a7-f088396830f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077903 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5530de4-963e-4282-9e46-5ead5330121f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077916 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqcbx\" (UniqueName: \"kubernetes.io/projected/b0640587-71f4-44e4-a6a5-5dc267d57844-kube-api-access-lqcbx\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077929 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfhq5\" (UniqueName: \"kubernetes.io/projected/6bfd336f-9602-43a5-b18e-2f7d24005d5a-kube-api-access-zfhq5\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.077937 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x9t2\" (UniqueName: \"kubernetes.io/projected/c5530de4-963e-4282-9e46-5ead5330121f-kube-api-access-4x9t2\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.081290 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-kube-api-access-mgbz9" (OuterVolumeSpecName: "kube-api-access-mgbz9") pod "57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" (UID: "57ac644c-0670-49bb-b7ce-3d9ecb72fa5a"). InnerVolumeSpecName "kube-api-access-mgbz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.185452 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgbz9\" (UniqueName: \"kubernetes.io/projected/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a-kube-api-access-mgbz9\") on node \"crc\" DevicePath \"\"" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.186793 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8lddn"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.190412 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8lddn"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.203532 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m92l5"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.206759 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m92l5"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.221221 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r46zk"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.892681 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" event={"ID":"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88","Type":"ContainerStarted","Data":"3287ef970e327a488035561f729d530303aeb6da225c9dc736774119ba3805a2"} Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.893166 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" event={"ID":"4d6cbfa1-a2bb-4c11-87b2-a5897573ed88","Type":"ContainerStarted","Data":"2cf9813ba9fac29c2e3612d6f63d46e9b57e46abe95849bc62dc89479a7d2f40"} Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.893206 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.895717 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.898413 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" event={"ID":"57ac644c-0670-49bb-b7ce-3d9ecb72fa5a","Type":"ContainerDied","Data":"aaa1e6313cc354ad0aba9c7300c99e256df61561199e2e7718665afe80523bbb"} Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.898462 4987 scope.go:117] "RemoveContainer" containerID="b7acd789d89c4ef81a64ca88670dbbbce5fa6aaca5b5c3b741b29cfb477b1c5b" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.898506 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rzqt7" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.898525 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-98fbt" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.898742 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-j9wm6" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.917776 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r46zk" podStartSLOduration=1.91775267 podStartE2EDuration="1.91775267s" podCreationTimestamp="2025-10-04 07:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:14:08.912303841 +0000 UTC m=+265.967202683" watchObservedRunningTime="2025-10-04 07:14:08.91775267 +0000 UTC m=+265.972651512" Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.928981 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rzqt7"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.935325 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rzqt7"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.944755 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-98fbt"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.952503 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-98fbt"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.977170 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9wm6"] Oct 04 07:14:08 crc kubenswrapper[4987]: I1004 07:14:08.979591 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-j9wm6"] Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551112 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jq5sw"] Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551295 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551308 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551318 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551324 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551332 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551338 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551350 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551356 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551368 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551373 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551380 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551385 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551393 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerName="marketplace-operator" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551399 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerName="marketplace-operator" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551408 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551414 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551422 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551427 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551435 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551442 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551456 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551465 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551477 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551485 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="extract-content" Oct 04 07:14:09 crc kubenswrapper[4987]: E1004 07:14:09.551494 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551500 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="extract-utilities" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551583 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551596 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5530de4-963e-4282-9e46-5ead5330121f" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551605 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" containerName="marketplace-operator" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.551614 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.554674 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" containerName="registry-server" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.555364 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.578137 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.583355 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq5sw"] Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.607395 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s8f7\" (UniqueName: \"kubernetes.io/projected/c5c87be9-03bb-486d-8024-8f88ee8f8384-kube-api-access-2s8f7\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.607532 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5c87be9-03bb-486d-8024-8f88ee8f8384-utilities\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.607603 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5c87be9-03bb-486d-8024-8f88ee8f8384-catalog-content\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.708944 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s8f7\" (UniqueName: \"kubernetes.io/projected/c5c87be9-03bb-486d-8024-8f88ee8f8384-kube-api-access-2s8f7\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.709053 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5c87be9-03bb-486d-8024-8f88ee8f8384-utilities\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.709078 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5c87be9-03bb-486d-8024-8f88ee8f8384-catalog-content\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.709535 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5c87be9-03bb-486d-8024-8f88ee8f8384-catalog-content\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.709699 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5c87be9-03bb-486d-8024-8f88ee8f8384-utilities\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.733465 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s8f7\" (UniqueName: \"kubernetes.io/projected/c5c87be9-03bb-486d-8024-8f88ee8f8384-kube-api-access-2s8f7\") pod \"redhat-marketplace-jq5sw\" (UID: \"c5c87be9-03bb-486d-8024-8f88ee8f8384\") " pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.766986 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rj2zj"] Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.768169 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.771193 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.779004 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rj2zj"] Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.809993 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deff79f-24f7-42c9-9485-b846b87ada21-catalog-content\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.810065 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksvwn\" (UniqueName: \"kubernetes.io/projected/1deff79f-24f7-42c9-9485-b846b87ada21-kube-api-access-ksvwn\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.810099 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deff79f-24f7-42c9-9485-b846b87ada21-utilities\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.871078 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.911766 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksvwn\" (UniqueName: \"kubernetes.io/projected/1deff79f-24f7-42c9-9485-b846b87ada21-kube-api-access-ksvwn\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.912470 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deff79f-24f7-42c9-9485-b846b87ada21-utilities\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.912548 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deff79f-24f7-42c9-9485-b846b87ada21-catalog-content\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.913188 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1deff79f-24f7-42c9-9485-b846b87ada21-catalog-content\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.913408 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1deff79f-24f7-42c9-9485-b846b87ada21-utilities\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:09 crc kubenswrapper[4987]: I1004 07:14:09.938266 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksvwn\" (UniqueName: \"kubernetes.io/projected/1deff79f-24f7-42c9-9485-b846b87ada21-kube-api-access-ksvwn\") pod \"redhat-operators-rj2zj\" (UID: \"1deff79f-24f7-42c9-9485-b846b87ada21\") " pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.061784 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2841f4fc-7a32-4d8a-91a7-f088396830f3" path="/var/lib/kubelet/pods/2841f4fc-7a32-4d8a-91a7-f088396830f3/volumes" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.062616 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57ac644c-0670-49bb-b7ce-3d9ecb72fa5a" path="/var/lib/kubelet/pods/57ac644c-0670-49bb-b7ce-3d9ecb72fa5a/volumes" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.063131 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bfd336f-9602-43a5-b18e-2f7d24005d5a" path="/var/lib/kubelet/pods/6bfd336f-9602-43a5-b18e-2f7d24005d5a/volumes" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.064149 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0640587-71f4-44e4-a6a5-5dc267d57844" path="/var/lib/kubelet/pods/b0640587-71f4-44e4-a6a5-5dc267d57844/volumes" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.064966 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5530de4-963e-4282-9e46-5ead5330121f" path="/var/lib/kubelet/pods/c5530de4-963e-4282-9e46-5ead5330121f/volumes" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.088285 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.291521 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jq5sw"] Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.484382 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rj2zj"] Oct 04 07:14:10 crc kubenswrapper[4987]: W1004 07:14:10.524678 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1deff79f_24f7_42c9_9485_b846b87ada21.slice/crio-91b73a543cc2287ce7abf6e3f0ba7e975aaa62090f34fc7e0fa0172f4773f1da WatchSource:0}: Error finding container 91b73a543cc2287ce7abf6e3f0ba7e975aaa62090f34fc7e0fa0172f4773f1da: Status 404 returned error can't find the container with id 91b73a543cc2287ce7abf6e3f0ba7e975aaa62090f34fc7e0fa0172f4773f1da Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.912451 4987 generic.go:334] "Generic (PLEG): container finished" podID="1deff79f-24f7-42c9-9485-b846b87ada21" containerID="5fad7d8d3d1654e0fbf22c72751e1b7e52639280bb49d453d53672c34a28d38f" exitCode=0 Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.912549 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2zj" event={"ID":"1deff79f-24f7-42c9-9485-b846b87ada21","Type":"ContainerDied","Data":"5fad7d8d3d1654e0fbf22c72751e1b7e52639280bb49d453d53672c34a28d38f"} Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.912614 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2zj" event={"ID":"1deff79f-24f7-42c9-9485-b846b87ada21","Type":"ContainerStarted","Data":"91b73a543cc2287ce7abf6e3f0ba7e975aaa62090f34fc7e0fa0172f4773f1da"} Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.913928 4987 generic.go:334] "Generic (PLEG): container finished" podID="c5c87be9-03bb-486d-8024-8f88ee8f8384" containerID="0309da8bc601be41468bb34f76015ec57925000bd56df7c46de73017095e6887" exitCode=0 Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.913982 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq5sw" event={"ID":"c5c87be9-03bb-486d-8024-8f88ee8f8384","Type":"ContainerDied","Data":"0309da8bc601be41468bb34f76015ec57925000bd56df7c46de73017095e6887"} Oct 04 07:14:10 crc kubenswrapper[4987]: I1004 07:14:10.914007 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq5sw" event={"ID":"c5c87be9-03bb-486d-8024-8f88ee8f8384","Type":"ContainerStarted","Data":"1bdbb6719a94fe2ce17794cd84430c566c62afc0333d61af9615e5855ef75939"} Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.920469 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2zj" event={"ID":"1deff79f-24f7-42c9-9485-b846b87ada21","Type":"ContainerStarted","Data":"a53bf73cf32cbbffc90e560c3b0a4ce3288ae57340d8bba70d2b8fff701759c7"} Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.925269 4987 generic.go:334] "Generic (PLEG): container finished" podID="c5c87be9-03bb-486d-8024-8f88ee8f8384" containerID="9e2d0522bac1733e9908acc3a68672978c1d8f6c8423b029856e99fa121aa9d1" exitCode=0 Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.925326 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq5sw" event={"ID":"c5c87be9-03bb-486d-8024-8f88ee8f8384","Type":"ContainerDied","Data":"9e2d0522bac1733e9908acc3a68672978c1d8f6c8423b029856e99fa121aa9d1"} Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.949417 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2mw6d"] Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.955721 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.957932 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 07:14:11 crc kubenswrapper[4987]: I1004 07:14:11.958731 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2mw6d"] Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.041463 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-catalog-content\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.041565 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d57nh\" (UniqueName: \"kubernetes.io/projected/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-kube-api-access-d57nh\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.041611 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-utilities\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.137142 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rmvdf"] Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.138297 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.140532 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.142704 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-catalog-content\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.142817 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d57nh\" (UniqueName: \"kubernetes.io/projected/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-kube-api-access-d57nh\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.142877 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-utilities\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.143122 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-catalog-content\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.143401 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-utilities\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.170212 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d57nh\" (UniqueName: \"kubernetes.io/projected/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-kube-api-access-d57nh\") pod \"community-operators-2mw6d\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.181337 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmvdf"] Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.244284 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb226448-ba0b-4cf6-9000-1f3390408256-catalog-content\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.244345 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb226448-ba0b-4cf6-9000-1f3390408256-utilities\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.244389 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxs2t\" (UniqueName: \"kubernetes.io/projected/cb226448-ba0b-4cf6-9000-1f3390408256-kube-api-access-sxs2t\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.287693 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.345368 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb226448-ba0b-4cf6-9000-1f3390408256-catalog-content\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.345618 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb226448-ba0b-4cf6-9000-1f3390408256-utilities\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.345726 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxs2t\" (UniqueName: \"kubernetes.io/projected/cb226448-ba0b-4cf6-9000-1f3390408256-kube-api-access-sxs2t\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.345821 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb226448-ba0b-4cf6-9000-1f3390408256-catalog-content\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.346069 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb226448-ba0b-4cf6-9000-1f3390408256-utilities\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.364537 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxs2t\" (UniqueName: \"kubernetes.io/projected/cb226448-ba0b-4cf6-9000-1f3390408256-kube-api-access-sxs2t\") pod \"certified-operators-rmvdf\" (UID: \"cb226448-ba0b-4cf6-9000-1f3390408256\") " pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.450753 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.685294 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2mw6d"] Oct 04 07:14:12 crc kubenswrapper[4987]: W1004 07:14:12.696585 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f3bd551_83fb_4b4c_98d3_a96dc30119ee.slice/crio-dab9e86a16b835bf41ca09717538f945dbbf70508790749742352850f8444cf4 WatchSource:0}: Error finding container dab9e86a16b835bf41ca09717538f945dbbf70508790749742352850f8444cf4: Status 404 returned error can't find the container with id dab9e86a16b835bf41ca09717538f945dbbf70508790749742352850f8444cf4 Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.821130 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmvdf"] Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.936397 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jq5sw" event={"ID":"c5c87be9-03bb-486d-8024-8f88ee8f8384","Type":"ContainerStarted","Data":"c3f3e44f4a0117a012f85be76ab9c649571051287ed8920e824f34016875d5a1"} Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.939678 4987 generic.go:334] "Generic (PLEG): container finished" podID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerID="cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629" exitCode=0 Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.939781 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mw6d" event={"ID":"5f3bd551-83fb-4b4c-98d3-a96dc30119ee","Type":"ContainerDied","Data":"cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629"} Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.939841 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mw6d" event={"ID":"5f3bd551-83fb-4b4c-98d3-a96dc30119ee","Type":"ContainerStarted","Data":"dab9e86a16b835bf41ca09717538f945dbbf70508790749742352850f8444cf4"} Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.941832 4987 generic.go:334] "Generic (PLEG): container finished" podID="1deff79f-24f7-42c9-9485-b846b87ada21" containerID="a53bf73cf32cbbffc90e560c3b0a4ce3288ae57340d8bba70d2b8fff701759c7" exitCode=0 Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.941873 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2zj" event={"ID":"1deff79f-24f7-42c9-9485-b846b87ada21","Type":"ContainerDied","Data":"a53bf73cf32cbbffc90e560c3b0a4ce3288ae57340d8bba70d2b8fff701759c7"} Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.944661 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmvdf" event={"ID":"cb226448-ba0b-4cf6-9000-1f3390408256","Type":"ContainerStarted","Data":"492ed412f7341c0815756e3ddd573b76305a7038a2e640848fa2ce1d23de473d"} Oct 04 07:14:12 crc kubenswrapper[4987]: I1004 07:14:12.957145 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jq5sw" podStartSLOduration=2.244980868 podStartE2EDuration="3.95712801s" podCreationTimestamp="2025-10-04 07:14:09 +0000 UTC" firstStartedPulling="2025-10-04 07:14:10.91967613 +0000 UTC m=+267.974574972" lastFinishedPulling="2025-10-04 07:14:12.631823272 +0000 UTC m=+269.686722114" observedRunningTime="2025-10-04 07:14:12.954836059 +0000 UTC m=+270.009734901" watchObservedRunningTime="2025-10-04 07:14:12.95712801 +0000 UTC m=+270.012026852" Oct 04 07:14:13 crc kubenswrapper[4987]: I1004 07:14:13.951277 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2zj" event={"ID":"1deff79f-24f7-42c9-9485-b846b87ada21","Type":"ContainerStarted","Data":"41cb406faa9ae9a10a19f9c079ce68bfba400b880983ba5eb8e460cff6b127cb"} Oct 04 07:14:13 crc kubenswrapper[4987]: I1004 07:14:13.953413 4987 generic.go:334] "Generic (PLEG): container finished" podID="cb226448-ba0b-4cf6-9000-1f3390408256" containerID="73414dded4e0d760d48f36a483eb4445783b71a46763420354f1cefeec9f8088" exitCode=0 Oct 04 07:14:13 crc kubenswrapper[4987]: I1004 07:14:13.953510 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmvdf" event={"ID":"cb226448-ba0b-4cf6-9000-1f3390408256","Type":"ContainerDied","Data":"73414dded4e0d760d48f36a483eb4445783b71a46763420354f1cefeec9f8088"} Oct 04 07:14:13 crc kubenswrapper[4987]: I1004 07:14:13.973912 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rj2zj" podStartSLOduration=2.512659894 podStartE2EDuration="4.973884339s" podCreationTimestamp="2025-10-04 07:14:09 +0000 UTC" firstStartedPulling="2025-10-04 07:14:10.916195983 +0000 UTC m=+267.971094825" lastFinishedPulling="2025-10-04 07:14:13.377420438 +0000 UTC m=+270.432319270" observedRunningTime="2025-10-04 07:14:13.971818585 +0000 UTC m=+271.026717467" watchObservedRunningTime="2025-10-04 07:14:13.973884339 +0000 UTC m=+271.028783201" Oct 04 07:14:14 crc kubenswrapper[4987]: I1004 07:14:14.961640 4987 generic.go:334] "Generic (PLEG): container finished" podID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerID="bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e" exitCode=0 Oct 04 07:14:14 crc kubenswrapper[4987]: I1004 07:14:14.961729 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mw6d" event={"ID":"5f3bd551-83fb-4b4c-98d3-a96dc30119ee","Type":"ContainerDied","Data":"bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e"} Oct 04 07:14:15 crc kubenswrapper[4987]: I1004 07:14:15.970470 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmvdf" event={"ID":"cb226448-ba0b-4cf6-9000-1f3390408256","Type":"ContainerStarted","Data":"86bd8373a04fa10491a7463cfe75a18695a49617f81afd7ccf63e35b40ecdfda"} Oct 04 07:14:15 crc kubenswrapper[4987]: I1004 07:14:15.973887 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mw6d" event={"ID":"5f3bd551-83fb-4b4c-98d3-a96dc30119ee","Type":"ContainerStarted","Data":"ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0"} Oct 04 07:14:16 crc kubenswrapper[4987]: I1004 07:14:16.008101 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2mw6d" podStartSLOduration=2.315570353 podStartE2EDuration="5.008083908s" podCreationTimestamp="2025-10-04 07:14:11 +0000 UTC" firstStartedPulling="2025-10-04 07:14:12.941244318 +0000 UTC m=+269.996143160" lastFinishedPulling="2025-10-04 07:14:15.633757873 +0000 UTC m=+272.688656715" observedRunningTime="2025-10-04 07:14:16.006839369 +0000 UTC m=+273.061738211" watchObservedRunningTime="2025-10-04 07:14:16.008083908 +0000 UTC m=+273.062982750" Oct 04 07:14:16 crc kubenswrapper[4987]: I1004 07:14:16.983125 4987 generic.go:334] "Generic (PLEG): container finished" podID="cb226448-ba0b-4cf6-9000-1f3390408256" containerID="86bd8373a04fa10491a7463cfe75a18695a49617f81afd7ccf63e35b40ecdfda" exitCode=0 Oct 04 07:14:16 crc kubenswrapper[4987]: I1004 07:14:16.983183 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmvdf" event={"ID":"cb226448-ba0b-4cf6-9000-1f3390408256","Type":"ContainerDied","Data":"86bd8373a04fa10491a7463cfe75a18695a49617f81afd7ccf63e35b40ecdfda"} Oct 04 07:14:17 crc kubenswrapper[4987]: I1004 07:14:17.991283 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmvdf" event={"ID":"cb226448-ba0b-4cf6-9000-1f3390408256","Type":"ContainerStarted","Data":"c08bb2e31c5927a09c1c9b504765c2f3c4a292f78c471aee725082487f640a95"} Oct 04 07:14:18 crc kubenswrapper[4987]: I1004 07:14:18.019015 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rmvdf" podStartSLOduration=2.224891707 podStartE2EDuration="6.018988386s" podCreationTimestamp="2025-10-04 07:14:12 +0000 UTC" firstStartedPulling="2025-10-04 07:14:13.96002609 +0000 UTC m=+271.014924932" lastFinishedPulling="2025-10-04 07:14:17.754122769 +0000 UTC m=+274.809021611" observedRunningTime="2025-10-04 07:14:18.010220815 +0000 UTC m=+275.065119657" watchObservedRunningTime="2025-10-04 07:14:18.018988386 +0000 UTC m=+275.073887238" Oct 04 07:14:19 crc kubenswrapper[4987]: I1004 07:14:19.872264 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:19 crc kubenswrapper[4987]: I1004 07:14:19.872572 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:19 crc kubenswrapper[4987]: I1004 07:14:19.914951 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:20 crc kubenswrapper[4987]: I1004 07:14:20.035909 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jq5sw" Oct 04 07:14:20 crc kubenswrapper[4987]: I1004 07:14:20.089125 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:20 crc kubenswrapper[4987]: I1004 07:14:20.089179 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:20 crc kubenswrapper[4987]: I1004 07:14:20.126186 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:21 crc kubenswrapper[4987]: I1004 07:14:21.047124 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rj2zj" Oct 04 07:14:22 crc kubenswrapper[4987]: I1004 07:14:22.287847 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:22 crc kubenswrapper[4987]: I1004 07:14:22.288240 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:22 crc kubenswrapper[4987]: I1004 07:14:22.324505 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:22 crc kubenswrapper[4987]: I1004 07:14:22.451711 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:22 crc kubenswrapper[4987]: I1004 07:14:22.451812 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:22 crc kubenswrapper[4987]: I1004 07:14:22.488767 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:14:23 crc kubenswrapper[4987]: I1004 07:14:23.058602 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:14:23 crc kubenswrapper[4987]: I1004 07:14:23.066853 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rmvdf" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.158379 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z"] Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.159517 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.165361 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.165386 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.184183 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z"] Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.281963 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d670748a-02a1-47cb-bcb5-7c8015ee7107-secret-volume\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.282016 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d670748a-02a1-47cb-bcb5-7c8015ee7107-config-volume\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.282054 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8m29\" (UniqueName: \"kubernetes.io/projected/d670748a-02a1-47cb-bcb5-7c8015ee7107-kube-api-access-c8m29\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.383450 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d670748a-02a1-47cb-bcb5-7c8015ee7107-config-volume\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.383492 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d670748a-02a1-47cb-bcb5-7c8015ee7107-secret-volume\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.383525 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8m29\" (UniqueName: \"kubernetes.io/projected/d670748a-02a1-47cb-bcb5-7c8015ee7107-kube-api-access-c8m29\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.384547 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d670748a-02a1-47cb-bcb5-7c8015ee7107-config-volume\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.395286 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d670748a-02a1-47cb-bcb5-7c8015ee7107-secret-volume\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.398732 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8m29\" (UniqueName: \"kubernetes.io/projected/d670748a-02a1-47cb-bcb5-7c8015ee7107-kube-api-access-c8m29\") pod \"collect-profiles-29326035-gj97z\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.476720 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:00 crc kubenswrapper[4987]: I1004 07:15:00.689792 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z"] Oct 04 07:15:01 crc kubenswrapper[4987]: I1004 07:15:01.235732 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" event={"ID":"d670748a-02a1-47cb-bcb5-7c8015ee7107","Type":"ContainerStarted","Data":"f79ea50b7e1dd8c68d658d03003fb54d0fff1beef4deae485f357b94fd45cd94"} Oct 04 07:15:01 crc kubenswrapper[4987]: I1004 07:15:01.235782 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" event={"ID":"d670748a-02a1-47cb-bcb5-7c8015ee7107","Type":"ContainerStarted","Data":"46f9c5e6fe09c01e44d00db9021a56b643346d8a853b6bb53b7d628ce0181521"} Oct 04 07:15:02 crc kubenswrapper[4987]: I1004 07:15:02.242269 4987 generic.go:334] "Generic (PLEG): container finished" podID="d670748a-02a1-47cb-bcb5-7c8015ee7107" containerID="f79ea50b7e1dd8c68d658d03003fb54d0fff1beef4deae485f357b94fd45cd94" exitCode=0 Oct 04 07:15:02 crc kubenswrapper[4987]: I1004 07:15:02.242310 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" event={"ID":"d670748a-02a1-47cb-bcb5-7c8015ee7107","Type":"ContainerDied","Data":"f79ea50b7e1dd8c68d658d03003fb54d0fff1beef4deae485f357b94fd45cd94"} Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.477669 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.521776 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d670748a-02a1-47cb-bcb5-7c8015ee7107-config-volume\") pod \"d670748a-02a1-47cb-bcb5-7c8015ee7107\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.521889 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d670748a-02a1-47cb-bcb5-7c8015ee7107-secret-volume\") pod \"d670748a-02a1-47cb-bcb5-7c8015ee7107\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.521913 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8m29\" (UniqueName: \"kubernetes.io/projected/d670748a-02a1-47cb-bcb5-7c8015ee7107-kube-api-access-c8m29\") pod \"d670748a-02a1-47cb-bcb5-7c8015ee7107\" (UID: \"d670748a-02a1-47cb-bcb5-7c8015ee7107\") " Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.522587 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d670748a-02a1-47cb-bcb5-7c8015ee7107-config-volume" (OuterVolumeSpecName: "config-volume") pod "d670748a-02a1-47cb-bcb5-7c8015ee7107" (UID: "d670748a-02a1-47cb-bcb5-7c8015ee7107"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.526752 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d670748a-02a1-47cb-bcb5-7c8015ee7107-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d670748a-02a1-47cb-bcb5-7c8015ee7107" (UID: "d670748a-02a1-47cb-bcb5-7c8015ee7107"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.526922 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d670748a-02a1-47cb-bcb5-7c8015ee7107-kube-api-access-c8m29" (OuterVolumeSpecName: "kube-api-access-c8m29") pod "d670748a-02a1-47cb-bcb5-7c8015ee7107" (UID: "d670748a-02a1-47cb-bcb5-7c8015ee7107"). InnerVolumeSpecName "kube-api-access-c8m29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.623051 4987 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d670748a-02a1-47cb-bcb5-7c8015ee7107-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.623091 4987 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d670748a-02a1-47cb-bcb5-7c8015ee7107-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:15:03 crc kubenswrapper[4987]: I1004 07:15:03.623102 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8m29\" (UniqueName: \"kubernetes.io/projected/d670748a-02a1-47cb-bcb5-7c8015ee7107-kube-api-access-c8m29\") on node \"crc\" DevicePath \"\"" Oct 04 07:15:04 crc kubenswrapper[4987]: I1004 07:15:04.253697 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" event={"ID":"d670748a-02a1-47cb-bcb5-7c8015ee7107","Type":"ContainerDied","Data":"46f9c5e6fe09c01e44d00db9021a56b643346d8a853b6bb53b7d628ce0181521"} Oct 04 07:15:04 crc kubenswrapper[4987]: I1004 07:15:04.253742 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46f9c5e6fe09c01e44d00db9021a56b643346d8a853b6bb53b7d628ce0181521" Oct 04 07:15:04 crc kubenswrapper[4987]: I1004 07:15:04.253743 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326035-gj97z" Oct 04 07:15:42 crc kubenswrapper[4987]: I1004 07:15:42.702898 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:15:42 crc kubenswrapper[4987]: I1004 07:15:42.703364 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:16:12 crc kubenswrapper[4987]: I1004 07:16:12.703180 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:16:12 crc kubenswrapper[4987]: I1004 07:16:12.703698 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:16:42 crc kubenswrapper[4987]: I1004 07:16:42.703025 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:16:42 crc kubenswrapper[4987]: I1004 07:16:42.703434 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:16:42 crc kubenswrapper[4987]: I1004 07:16:42.703483 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:16:42 crc kubenswrapper[4987]: I1004 07:16:42.704045 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1f7c32b490273e9ddd0a08c984ac824a51480ca48e9c2de0a6c24f34a688a2e"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:16:42 crc kubenswrapper[4987]: I1004 07:16:42.704102 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://d1f7c32b490273e9ddd0a08c984ac824a51480ca48e9c2de0a6c24f34a688a2e" gracePeriod=600 Oct 04 07:16:43 crc kubenswrapper[4987]: I1004 07:16:43.780732 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="d1f7c32b490273e9ddd0a08c984ac824a51480ca48e9c2de0a6c24f34a688a2e" exitCode=0 Oct 04 07:16:43 crc kubenswrapper[4987]: I1004 07:16:43.780800 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"d1f7c32b490273e9ddd0a08c984ac824a51480ca48e9c2de0a6c24f34a688a2e"} Oct 04 07:16:43 crc kubenswrapper[4987]: I1004 07:16:43.781273 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"505ec6340e253f2b31b71cdfb3e0b7658643e55c35199ae6c3e39c731aca0c48"} Oct 04 07:16:43 crc kubenswrapper[4987]: I1004 07:16:43.781298 4987 scope.go:117] "RemoveContainer" containerID="04038fd9f44a737738bd4f0776fee45d0991f48e595b58c4c93994155ac8f910" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.304375 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rhv7r"] Oct 04 07:17:17 crc kubenswrapper[4987]: E1004 07:17:17.305195 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d670748a-02a1-47cb-bcb5-7c8015ee7107" containerName="collect-profiles" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.305213 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="d670748a-02a1-47cb-bcb5-7c8015ee7107" containerName="collect-profiles" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.305330 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="d670748a-02a1-47cb-bcb5-7c8015ee7107" containerName="collect-profiles" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.305851 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.328664 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rhv7r"] Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456713 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456764 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456786 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-registry-certificates\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456803 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456834 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-registry-tls\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456862 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tw77\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-kube-api-access-6tw77\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456878 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-trusted-ca\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.456919 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-bound-sa-token\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.474812 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558404 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-bound-sa-token\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558468 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558491 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-registry-certificates\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558511 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558541 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-registry-tls\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558577 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tw77\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-kube-api-access-6tw77\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.558599 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-trusted-ca\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.559183 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.559980 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-trusted-ca\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.560062 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-registry-certificates\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.565049 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.565453 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-registry-tls\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.574469 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tw77\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-kube-api-access-6tw77\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.574885 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/feaf7f4f-38de-4312-98c3-c5ba40adcd0f-bound-sa-token\") pod \"image-registry-66df7c8f76-rhv7r\" (UID: \"feaf7f4f-38de-4312-98c3-c5ba40adcd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.624014 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.787061 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rhv7r"] Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.955226 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" event={"ID":"feaf7f4f-38de-4312-98c3-c5ba40adcd0f","Type":"ContainerStarted","Data":"11303a790b1b5a133b4917881012df122a63adde6d4a41f56e613f60fe156047"} Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.955275 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" event={"ID":"feaf7f4f-38de-4312-98c3-c5ba40adcd0f","Type":"ContainerStarted","Data":"a219f8f45db6f5ab29cb4e0811eb6efa548b919322ad9c4b35780b8869f0f20e"} Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.955590 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:17 crc kubenswrapper[4987]: I1004 07:17:17.975217 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" podStartSLOduration=0.975173535 podStartE2EDuration="975.173535ms" podCreationTimestamp="2025-10-04 07:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:17:17.973595914 +0000 UTC m=+455.028494776" watchObservedRunningTime="2025-10-04 07:17:17.975173535 +0000 UTC m=+455.030072377" Oct 04 07:17:37 crc kubenswrapper[4987]: I1004 07:17:37.628617 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rhv7r" Oct 04 07:17:37 crc kubenswrapper[4987]: I1004 07:17:37.687907 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mrxhg"] Oct 04 07:18:02 crc kubenswrapper[4987]: I1004 07:18:02.739384 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" podUID="001408b5-e454-432d-93dc-3a2dc497941f" containerName="registry" containerID="cri-o://83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8" gracePeriod=30 Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.103591 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.186884 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-registry-certificates\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.187273 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.187316 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm5rb\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-kube-api-access-dm5rb\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.187433 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/001408b5-e454-432d-93dc-3a2dc497941f-installation-pull-secrets\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.187481 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/001408b5-e454-432d-93dc-3a2dc497941f-ca-trust-extracted\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.187518 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-trusted-ca\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.188120 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.188262 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.192741 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-registry-tls\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.192857 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-bound-sa-token\") pod \"001408b5-e454-432d-93dc-3a2dc497941f\" (UID: \"001408b5-e454-432d-93dc-3a2dc497941f\") " Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.193588 4987 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.193617 4987 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/001408b5-e454-432d-93dc-3a2dc497941f-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.195366 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.195650 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001408b5-e454-432d-93dc-3a2dc497941f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.195673 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-kube-api-access-dm5rb" (OuterVolumeSpecName: "kube-api-access-dm5rb") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "kube-api-access-dm5rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.197375 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.207577 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/001408b5-e454-432d-93dc-3a2dc497941f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.208317 4987 generic.go:334] "Generic (PLEG): container finished" podID="001408b5-e454-432d-93dc-3a2dc497941f" containerID="83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8" exitCode=0 Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.208366 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" event={"ID":"001408b5-e454-432d-93dc-3a2dc497941f","Type":"ContainerDied","Data":"83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8"} Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.208396 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" event={"ID":"001408b5-e454-432d-93dc-3a2dc497941f","Type":"ContainerDied","Data":"4c55978594c509d2aa4b298fdc59a5cdd9768cc904a52ab9627ee016674e5012"} Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.208413 4987 scope.go:117] "RemoveContainer" containerID="83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.208535 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mrxhg" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.221199 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "001408b5-e454-432d-93dc-3a2dc497941f" (UID: "001408b5-e454-432d-93dc-3a2dc497941f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.237349 4987 scope.go:117] "RemoveContainer" containerID="83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8" Oct 04 07:18:03 crc kubenswrapper[4987]: E1004 07:18:03.237991 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8\": container with ID starting with 83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8 not found: ID does not exist" containerID="83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.238027 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8"} err="failed to get container status \"83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8\": rpc error: code = NotFound desc = could not find container \"83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8\": container with ID starting with 83569a8f071af5a1a244ea4df8f3a3700f694b2d41cf60927f36c9628be560e8 not found: ID does not exist" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.295249 4987 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/001408b5-e454-432d-93dc-3a2dc497941f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.295533 4987 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/001408b5-e454-432d-93dc-3a2dc497941f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.295545 4987 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.295554 4987 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.295564 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm5rb\" (UniqueName: \"kubernetes.io/projected/001408b5-e454-432d-93dc-3a2dc497941f-kube-api-access-dm5rb\") on node \"crc\" DevicePath \"\"" Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.535756 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mrxhg"] Oct 04 07:18:03 crc kubenswrapper[4987]: I1004 07:18:03.543337 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mrxhg"] Oct 04 07:18:04 crc kubenswrapper[4987]: I1004 07:18:04.060314 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001408b5-e454-432d-93dc-3a2dc497941f" path="/var/lib/kubelet/pods/001408b5-e454-432d-93dc-3a2dc497941f/volumes" Oct 04 07:18:42 crc kubenswrapper[4987]: I1004 07:18:42.702740 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:18:42 crc kubenswrapper[4987]: I1004 07:18:42.703343 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:18:44 crc kubenswrapper[4987]: I1004 07:18:44.216042 4987 scope.go:117] "RemoveContainer" containerID="0f0e301c86a98f19a06f0d9c527b53ecb2cc3688e836bd33038925bb79663c1e" Oct 04 07:18:44 crc kubenswrapper[4987]: I1004 07:18:44.237004 4987 scope.go:117] "RemoveContainer" containerID="519342791b021bbc00fe04ddc0784cd24fa3f662b598d017423d3367d01703ee" Oct 04 07:19:12 crc kubenswrapper[4987]: I1004 07:19:12.704100 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:19:12 crc kubenswrapper[4987]: I1004 07:19:12.705093 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:19:42 crc kubenswrapper[4987]: I1004 07:19:42.704036 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:19:42 crc kubenswrapper[4987]: I1004 07:19:42.704582 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:19:42 crc kubenswrapper[4987]: I1004 07:19:42.704644 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:19:42 crc kubenswrapper[4987]: I1004 07:19:42.705139 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"505ec6340e253f2b31b71cdfb3e0b7658643e55c35199ae6c3e39c731aca0c48"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:19:42 crc kubenswrapper[4987]: I1004 07:19:42.705190 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://505ec6340e253f2b31b71cdfb3e0b7658643e55c35199ae6c3e39c731aca0c48" gracePeriod=600 Oct 04 07:19:43 crc kubenswrapper[4987]: I1004 07:19:43.727694 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="505ec6340e253f2b31b71cdfb3e0b7658643e55c35199ae6c3e39c731aca0c48" exitCode=0 Oct 04 07:19:43 crc kubenswrapper[4987]: I1004 07:19:43.727738 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"505ec6340e253f2b31b71cdfb3e0b7658643e55c35199ae6c3e39c731aca0c48"} Oct 04 07:19:43 crc kubenswrapper[4987]: I1004 07:19:43.728032 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"241b883c886b42b50a0701b97bca81215e874c5bdbc8a5d84d217f37244c27ff"} Oct 04 07:19:43 crc kubenswrapper[4987]: I1004 07:19:43.728060 4987 scope.go:117] "RemoveContainer" containerID="d1f7c32b490273e9ddd0a08c984ac824a51480ca48e9c2de0a6c24f34a688a2e" Oct 04 07:19:44 crc kubenswrapper[4987]: I1004 07:19:44.271228 4987 scope.go:117] "RemoveContainer" containerID="54ea90b05288f109ced387a14e97c6464fe5f0d67d826fd6e99faf46dfc97652" Oct 04 07:19:44 crc kubenswrapper[4987]: I1004 07:19:44.294841 4987 scope.go:117] "RemoveContainer" containerID="e7bda8f7a5d29a75c07a5719638b10e2f51ece6f16031eb8853e9df796ca3bac" Oct 04 07:19:44 crc kubenswrapper[4987]: I1004 07:19:44.308914 4987 scope.go:117] "RemoveContainer" containerID="7be14b091da2f50792d72dbf8b344ae7ac96817713ffcd372b62c85470bd7bab" Oct 04 07:19:44 crc kubenswrapper[4987]: I1004 07:19:44.320530 4987 scope.go:117] "RemoveContainer" containerID="1db5992996b1526fea1643701e34958996048978e4464ec763e8188f991ce218" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.477578 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tw9rc"] Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482252 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-controller" containerID="cri-o://cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482718 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="sbdb" containerID="cri-o://def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482759 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="nbdb" containerID="cri-o://db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482789 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="northd" containerID="cri-o://e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482818 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482846 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-node" containerID="cri-o://ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.482876 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-acl-logging" containerID="cri-o://19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.535623 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" containerID="cri-o://d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" gracePeriod=30 Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.817758 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/3.log" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.819816 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovn-acl-logging/0.log" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.820211 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovn-controller/0.log" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.820646 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877010 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r2l62"] Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877281 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kubecfg-setup" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877296 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kubecfg-setup" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877304 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="sbdb" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877312 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="sbdb" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877327 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="nbdb" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877357 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="nbdb" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877370 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877377 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877386 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-acl-logging" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877394 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-acl-logging" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877402 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877413 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877452 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877461 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877470 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-node" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877477 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-node" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877487 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="northd" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877495 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="northd" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877524 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877531 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877541 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877546 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877553 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001408b5-e454-432d-93dc-3a2dc497941f" containerName="registry" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877560 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="001408b5-e454-432d-93dc-3a2dc497941f" containerName="registry" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877723 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-node" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877736 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877745 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877755 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877762 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877768 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877775 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="001408b5-e454-432d-93dc-3a2dc497941f" containerName="registry" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877781 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877787 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="northd" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877795 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovn-acl-logging" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877802 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877809 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="nbdb" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877816 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="sbdb" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877898 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877905 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: E1004 07:21:06.877914 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.877920 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerName="ovnkube-controller" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.879433 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997555 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-var-lib-openvswitch\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997609 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-log-socket\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997650 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-netd\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997663 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997674 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-netns\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997717 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997754 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-etc-openvswitch\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997776 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997798 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-systemd\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997802 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-log-socket" (OuterVolumeSpecName: "log-socket") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997830 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997857 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8rdv\" (UniqueName: \"kubernetes.io/projected/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-kube-api-access-k8rdv\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997887 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovn-node-metrics-cert\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997903 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-systemd-units\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997920 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-openvswitch\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997938 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-slash\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997950 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997966 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-kubelet\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997979 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-slash" (OuterVolumeSpecName: "host-slash") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.997982 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-env-overrides\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998017 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-ovn\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998005 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998049 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-bin\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998048 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998088 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-config\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998101 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998076 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998114 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-script-lib\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998129 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-node-log\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998162 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-ovn-kubernetes\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998189 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\" (UID: \"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3\") " Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998219 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-node-log" (OuterVolumeSpecName: "node-log") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998256 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998250 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998323 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998486 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-cni-bin\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998490 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998518 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-slash\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998530 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998548 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-node-log\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998589 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-log-socket\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998614 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-env-overrides\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998694 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-run-netns\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998728 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-etc-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998808 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-ovn\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998844 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998872 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-kubelet\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998890 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0ff11c21-a933-447d-ae2d-dc20953780a9-ovn-node-metrics-cert\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998908 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-systemd\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.998936 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999016 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-systemd-units\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999049 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999069 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-cni-netd\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999114 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-var-lib-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999135 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-ovnkube-config\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999170 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhxtr\" (UniqueName: \"kubernetes.io/projected/0ff11c21-a933-447d-ae2d-dc20953780a9-kube-api-access-xhxtr\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999203 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-ovnkube-script-lib\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999241 4987 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999251 4987 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999259 4987 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999267 4987 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999275 4987 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999283 4987 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999292 4987 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999301 4987 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999309 4987 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999317 4987 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999326 4987 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999334 4987 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999343 4987 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999351 4987 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999360 4987 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999367 4987 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:06 crc kubenswrapper[4987]: I1004 07:21:06.999374 4987 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.002756 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-kube-api-access-k8rdv" (OuterVolumeSpecName: "kube-api-access-k8rdv") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "kube-api-access-k8rdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.003436 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.010103 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" (UID: "c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.099972 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-slash\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100026 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-node-log\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100048 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-log-socket\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100069 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-env-overrides\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100091 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-run-netns\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100140 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-run-netns\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100158 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-slash\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100209 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-etc-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100215 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-log-socket\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100175 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-node-log\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100188 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-etc-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100295 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-ovn\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100327 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100372 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-kubelet\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100394 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0ff11c21-a933-447d-ae2d-dc20953780a9-ovn-node-metrics-cert\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100401 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-ovn\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100419 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-systemd\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100552 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100433 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-run-ovn-kubernetes\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100447 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-kubelet\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100445 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-systemd\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100655 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-systemd-units\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100654 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100693 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100714 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-cni-netd\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100758 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-run-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100770 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-systemd-units\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100801 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-cni-netd\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100775 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-env-overrides\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100785 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-var-lib-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100845 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-ovnkube-config\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100864 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhxtr\" (UniqueName: \"kubernetes.io/projected/0ff11c21-a933-447d-ae2d-dc20953780a9-kube-api-access-xhxtr\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100879 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-ovnkube-script-lib\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100905 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-cni-bin\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100937 4987 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100947 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8rdv\" (UniqueName: \"kubernetes.io/projected/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-kube-api-access-k8rdv\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100957 4987 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100980 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-host-cni-bin\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.100806 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/0ff11c21-a933-447d-ae2d-dc20953780a9-var-lib-openvswitch\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.101354 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-ovnkube-config\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.102758 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/0ff11c21-a933-447d-ae2d-dc20953780a9-ovnkube-script-lib\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.103459 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0ff11c21-a933-447d-ae2d-dc20953780a9-ovn-node-metrics-cert\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.116071 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhxtr\" (UniqueName: \"kubernetes.io/projected/0ff11c21-a933-447d-ae2d-dc20953780a9-kube-api-access-xhxtr\") pod \"ovnkube-node-r2l62\" (UID: \"0ff11c21-a933-447d-ae2d-dc20953780a9\") " pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.178609 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/2.log" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.179349 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/1.log" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.179455 4987 generic.go:334] "Generic (PLEG): container finished" podID="659797bb-8ec1-4ad8-880e-7aec3b068c46" containerID="1db87dc841e81f4e40e9a6c80e98c312d2b1ab807484d90d9f2e1be091bf01ad" exitCode=2 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.179531 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerDied","Data":"1db87dc841e81f4e40e9a6c80e98c312d2b1ab807484d90d9f2e1be091bf01ad"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.179592 4987 scope.go:117] "RemoveContainer" containerID="d2be8d2bb1132aaa5f4286f70ff406abf12275a633a8a0146cbefc05f920e942" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.180037 4987 scope.go:117] "RemoveContainer" containerID="1db87dc841e81f4e40e9a6c80e98c312d2b1ab807484d90d9f2e1be091bf01ad" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.180254 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mc8jj_openshift-multus(659797bb-8ec1-4ad8-880e-7aec3b068c46)\"" pod="openshift-multus/multus-mc8jj" podUID="659797bb-8ec1-4ad8-880e-7aec3b068c46" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.182748 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovnkube-controller/3.log" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.186416 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovn-acl-logging/0.log" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187329 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tw9rc_c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/ovn-controller/0.log" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187732 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" exitCode=0 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187765 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" exitCode=0 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187773 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" exitCode=0 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187780 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" exitCode=0 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187786 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" exitCode=0 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187792 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187810 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187838 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187851 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187860 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187870 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187879 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187889 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187899 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187905 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187910 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187915 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187921 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187925 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187930 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187935 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187940 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.187797 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" exitCode=0 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188021 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" exitCode=143 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188029 4987 generic.go:334] "Generic (PLEG): container finished" podID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" exitCode=143 Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188043 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188051 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188057 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188062 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188068 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188072 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188079 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188084 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188089 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188094 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188098 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188105 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188112 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188118 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188124 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188131 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188136 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188141 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188146 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188151 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188156 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188161 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188167 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tw9rc" event={"ID":"c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3","Type":"ContainerDied","Data":"f1084503219d9dab387d1525440270a1effad654748dbe8c4b08caa6def47c1f"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188174 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188181 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188186 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188191 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188195 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188201 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188207 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188212 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188217 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.188222 4987 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.191881 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.225399 4987 scope.go:117] "RemoveContainer" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.229106 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tw9rc"] Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.232301 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tw9rc"] Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.242279 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.260227 4987 scope.go:117] "RemoveContainer" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.275438 4987 scope.go:117] "RemoveContainer" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.288437 4987 scope.go:117] "RemoveContainer" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.325952 4987 scope.go:117] "RemoveContainer" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.338880 4987 scope.go:117] "RemoveContainer" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.353519 4987 scope.go:117] "RemoveContainer" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.364919 4987 scope.go:117] "RemoveContainer" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.436444 4987 scope.go:117] "RemoveContainer" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.451053 4987 scope.go:117] "RemoveContainer" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.451460 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": container with ID starting with d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2 not found: ID does not exist" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.451487 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} err="failed to get container status \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": rpc error: code = NotFound desc = could not find container \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": container with ID starting with d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.451512 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.451956 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": container with ID starting with 6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24 not found: ID does not exist" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.451976 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} err="failed to get container status \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": rpc error: code = NotFound desc = could not find container \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": container with ID starting with 6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.451988 4987 scope.go:117] "RemoveContainer" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.452215 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": container with ID starting with def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80 not found: ID does not exist" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.452236 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} err="failed to get container status \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": rpc error: code = NotFound desc = could not find container \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": container with ID starting with def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.452250 4987 scope.go:117] "RemoveContainer" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.452529 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": container with ID starting with db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac not found: ID does not exist" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.452556 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} err="failed to get container status \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": rpc error: code = NotFound desc = could not find container \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": container with ID starting with db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.452578 4987 scope.go:117] "RemoveContainer" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.452836 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": container with ID starting with e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca not found: ID does not exist" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.452857 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} err="failed to get container status \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": rpc error: code = NotFound desc = could not find container \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": container with ID starting with e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.452871 4987 scope.go:117] "RemoveContainer" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.453171 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": container with ID starting with d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272 not found: ID does not exist" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.453194 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} err="failed to get container status \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": rpc error: code = NotFound desc = could not find container \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": container with ID starting with d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.453211 4987 scope.go:117] "RemoveContainer" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.453406 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": container with ID starting with ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0 not found: ID does not exist" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.453423 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} err="failed to get container status \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": rpc error: code = NotFound desc = could not find container \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": container with ID starting with ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.453435 4987 scope.go:117] "RemoveContainer" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.453752 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": container with ID starting with 19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554 not found: ID does not exist" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.453771 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} err="failed to get container status \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": rpc error: code = NotFound desc = could not find container \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": container with ID starting with 19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.453784 4987 scope.go:117] "RemoveContainer" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.454019 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": container with ID starting with cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea not found: ID does not exist" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.454040 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} err="failed to get container status \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": rpc error: code = NotFound desc = could not find container \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": container with ID starting with cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.454054 4987 scope.go:117] "RemoveContainer" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" Oct 04 07:21:07 crc kubenswrapper[4987]: E1004 07:21:07.454454 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": container with ID starting with eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24 not found: ID does not exist" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.454472 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} err="failed to get container status \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": rpc error: code = NotFound desc = could not find container \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": container with ID starting with eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.454484 4987 scope.go:117] "RemoveContainer" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.454708 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} err="failed to get container status \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": rpc error: code = NotFound desc = could not find container \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": container with ID starting with d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.454733 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455021 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} err="failed to get container status \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": rpc error: code = NotFound desc = could not find container \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": container with ID starting with 6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455038 4987 scope.go:117] "RemoveContainer" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455267 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} err="failed to get container status \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": rpc error: code = NotFound desc = could not find container \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": container with ID starting with def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455295 4987 scope.go:117] "RemoveContainer" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455515 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} err="failed to get container status \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": rpc error: code = NotFound desc = could not find container \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": container with ID starting with db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455532 4987 scope.go:117] "RemoveContainer" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455765 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} err="failed to get container status \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": rpc error: code = NotFound desc = could not find container \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": container with ID starting with e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.455787 4987 scope.go:117] "RemoveContainer" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456053 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} err="failed to get container status \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": rpc error: code = NotFound desc = could not find container \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": container with ID starting with d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456068 4987 scope.go:117] "RemoveContainer" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456272 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} err="failed to get container status \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": rpc error: code = NotFound desc = could not find container \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": container with ID starting with ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456292 4987 scope.go:117] "RemoveContainer" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456490 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} err="failed to get container status \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": rpc error: code = NotFound desc = could not find container \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": container with ID starting with 19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456505 4987 scope.go:117] "RemoveContainer" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456804 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} err="failed to get container status \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": rpc error: code = NotFound desc = could not find container \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": container with ID starting with cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.456826 4987 scope.go:117] "RemoveContainer" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.457188 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} err="failed to get container status \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": rpc error: code = NotFound desc = could not find container \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": container with ID starting with eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.457260 4987 scope.go:117] "RemoveContainer" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.457530 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} err="failed to get container status \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": rpc error: code = NotFound desc = could not find container \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": container with ID starting with d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.457549 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.457763 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} err="failed to get container status \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": rpc error: code = NotFound desc = could not find container \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": container with ID starting with 6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.457786 4987 scope.go:117] "RemoveContainer" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.458008 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} err="failed to get container status \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": rpc error: code = NotFound desc = could not find container \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": container with ID starting with def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.458026 4987 scope.go:117] "RemoveContainer" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.458256 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} err="failed to get container status \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": rpc error: code = NotFound desc = could not find container \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": container with ID starting with db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.458282 4987 scope.go:117] "RemoveContainer" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.458712 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} err="failed to get container status \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": rpc error: code = NotFound desc = could not find container \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": container with ID starting with e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.458731 4987 scope.go:117] "RemoveContainer" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459012 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} err="failed to get container status \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": rpc error: code = NotFound desc = could not find container \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": container with ID starting with d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459033 4987 scope.go:117] "RemoveContainer" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459292 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} err="failed to get container status \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": rpc error: code = NotFound desc = could not find container \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": container with ID starting with ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459314 4987 scope.go:117] "RemoveContainer" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459649 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} err="failed to get container status \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": rpc error: code = NotFound desc = could not find container \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": container with ID starting with 19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459668 4987 scope.go:117] "RemoveContainer" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459868 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} err="failed to get container status \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": rpc error: code = NotFound desc = could not find container \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": container with ID starting with cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.459887 4987 scope.go:117] "RemoveContainer" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460076 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} err="failed to get container status \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": rpc error: code = NotFound desc = could not find container \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": container with ID starting with eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460091 4987 scope.go:117] "RemoveContainer" containerID="d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460307 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2"} err="failed to get container status \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": rpc error: code = NotFound desc = could not find container \"d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2\": container with ID starting with d345edf5aca20a599b2d5f5d22456135e4a4d76c9a94adb36889896c074ae9e2 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460322 4987 scope.go:117] "RemoveContainer" containerID="6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460564 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24"} err="failed to get container status \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": rpc error: code = NotFound desc = could not find container \"6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24\": container with ID starting with 6ec18ee51d2898c830ab163d3b8e7cdc56f84232910aa2cdb5a4c46880ff2c24 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460583 4987 scope.go:117] "RemoveContainer" containerID="def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460791 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80"} err="failed to get container status \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": rpc error: code = NotFound desc = could not find container \"def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80\": container with ID starting with def53ab9b3006a836fb5533d734519c77f54a6beff61027698b1dcc4559a4d80 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.460808 4987 scope.go:117] "RemoveContainer" containerID="db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461000 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac"} err="failed to get container status \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": rpc error: code = NotFound desc = could not find container \"db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac\": container with ID starting with db1b33d035434d2b81870a11eb2aafcc32fe49e04b9783d1cc475a2aa55e50ac not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461015 4987 scope.go:117] "RemoveContainer" containerID="e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461209 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca"} err="failed to get container status \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": rpc error: code = NotFound desc = could not find container \"e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca\": container with ID starting with e607bef6e04d2f2bbf1c1c663368b0c31c733080cda6327a9e4fdfb4436195ca not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461225 4987 scope.go:117] "RemoveContainer" containerID="d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461397 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272"} err="failed to get container status \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": rpc error: code = NotFound desc = could not find container \"d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272\": container with ID starting with d03a5d33d40d4cff984cf2953c93ca8d388ee592830427cb5ea93e6825b13272 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461413 4987 scope.go:117] "RemoveContainer" containerID="ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461732 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0"} err="failed to get container status \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": rpc error: code = NotFound desc = could not find container \"ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0\": container with ID starting with ae537cf7860b5a4b84ac2a7f88bcdae83bd0acdb7a1c3bd0b31d7b143ed373b0 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461755 4987 scope.go:117] "RemoveContainer" containerID="19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461951 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554"} err="failed to get container status \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": rpc error: code = NotFound desc = could not find container \"19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554\": container with ID starting with 19c1753cb2ff1a40937a8ec88dc070aeaba219217742fe1cb2313d095ad01554 not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.461970 4987 scope.go:117] "RemoveContainer" containerID="cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.462171 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea"} err="failed to get container status \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": rpc error: code = NotFound desc = could not find container \"cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea\": container with ID starting with cef7141da40c67915dc3ea4708b662744e8564f0c588178db25cf27d94b877ea not found: ID does not exist" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.462191 4987 scope.go:117] "RemoveContainer" containerID="eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24" Oct 04 07:21:07 crc kubenswrapper[4987]: I1004 07:21:07.462344 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24"} err="failed to get container status \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": rpc error: code = NotFound desc = could not find container \"eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24\": container with ID starting with eca8c7e10b295d1f81411fa2e12a0ed3392b98c178d116f68f75e19ff9bf9a24 not found: ID does not exist" Oct 04 07:21:08 crc kubenswrapper[4987]: I1004 07:21:08.062037 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3" path="/var/lib/kubelet/pods/c8b42ffd-33d6-48ad-9a54-1bfad0c6bba3/volumes" Oct 04 07:21:08 crc kubenswrapper[4987]: I1004 07:21:08.194818 4987 generic.go:334] "Generic (PLEG): container finished" podID="0ff11c21-a933-447d-ae2d-dc20953780a9" containerID="01d7b8cb22bb417517268357f89bcc1eb1d1bfe7941ed614fe989f66255bbdc7" exitCode=0 Oct 04 07:21:08 crc kubenswrapper[4987]: I1004 07:21:08.195080 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerDied","Data":"01d7b8cb22bb417517268357f89bcc1eb1d1bfe7941ed614fe989f66255bbdc7"} Oct 04 07:21:08 crc kubenswrapper[4987]: I1004 07:21:08.195198 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"d8e6284632421d7ef547d99b0e0e89b2dfd68556a098c8524fa4975083afd5b9"} Oct 04 07:21:08 crc kubenswrapper[4987]: I1004 07:21:08.199130 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/2.log" Oct 04 07:21:09 crc kubenswrapper[4987]: I1004 07:21:09.207104 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"ee1965c56fc0845d46d838487543eca793ee69f4aae0468d67229608155be4b9"} Oct 04 07:21:09 crc kubenswrapper[4987]: I1004 07:21:09.207602 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"1562b6e5397a5c7af619b00b21e13446b7ff525ffdddd6a0c0c8889e1da6c9e3"} Oct 04 07:21:09 crc kubenswrapper[4987]: I1004 07:21:09.207616 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"e8bf3933f39ac4aa9cc0388b06c955cc5fcfc3f9d4273a464d34681331a96553"} Oct 04 07:21:09 crc kubenswrapper[4987]: I1004 07:21:09.207643 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"3e5db3fec6ea939d60d7348e67d2fff93f39b3ded2c4afcf90569cbe1f2fdc25"} Oct 04 07:21:09 crc kubenswrapper[4987]: I1004 07:21:09.207653 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"c5852a7b647abda32930af36f3d04be9a0af98bd190f2229f818d5bae2b354f4"} Oct 04 07:21:09 crc kubenswrapper[4987]: I1004 07:21:09.207662 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"d62d02691618823b5bbf4eceeb259676039b857adcbe32056ebc8e62f3730a99"} Oct 04 07:21:11 crc kubenswrapper[4987]: I1004 07:21:11.221564 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"554a70dc2198d01727e390277814cebf3ef30fe9d1c7d69f35994a9c7c85bb0b"} Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.239574 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" event={"ID":"0ff11c21-a933-447d-ae2d-dc20953780a9","Type":"ContainerStarted","Data":"ba32de882fbe5ea424b7aa4990efdec538a4d4542654f57c537dfa2f9bda9f09"} Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.240157 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.240171 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.240181 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.277560 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.281517 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" podStartSLOduration=8.281498848 podStartE2EDuration="8.281498848s" podCreationTimestamp="2025-10-04 07:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:21:14.278858096 +0000 UTC m=+691.333756938" watchObservedRunningTime="2025-10-04 07:21:14.281498848 +0000 UTC m=+691.336397690" Oct 04 07:21:14 crc kubenswrapper[4987]: I1004 07:21:14.288794 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.479801 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn"] Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.480983 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.483072 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.493210 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn"] Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.613915 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.614071 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skps2\" (UniqueName: \"kubernetes.io/projected/92de2dd9-3957-4f74-a21e-ed5b040b5de4-kube-api-access-skps2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.614141 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.716267 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skps2\" (UniqueName: \"kubernetes.io/projected/92de2dd9-3957-4f74-a21e-ed5b040b5de4-kube-api-access-skps2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.716326 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.716393 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.716933 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.716943 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.738326 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skps2\" (UniqueName: \"kubernetes.io/projected/92de2dd9-3957-4f74-a21e-ed5b040b5de4-kube-api-access-skps2\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: I1004 07:21:15.795537 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: E1004 07:21:15.821108 4987 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(7de82a18fa572179fc0e1047bbe525e987f3b84b5b85ee5516634c3cb558ab0f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 07:21:15 crc kubenswrapper[4987]: E1004 07:21:15.821183 4987 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(7de82a18fa572179fc0e1047bbe525e987f3b84b5b85ee5516634c3cb558ab0f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: E1004 07:21:15.821209 4987 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(7de82a18fa572179fc0e1047bbe525e987f3b84b5b85ee5516634c3cb558ab0f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:15 crc kubenswrapper[4987]: E1004 07:21:15.821264 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace(92de2dd9-3957-4f74-a21e-ed5b040b5de4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace(92de2dd9-3957-4f74-a21e-ed5b040b5de4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(7de82a18fa572179fc0e1047bbe525e987f3b84b5b85ee5516634c3cb558ab0f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" Oct 04 07:21:16 crc kubenswrapper[4987]: I1004 07:21:16.249021 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:16 crc kubenswrapper[4987]: I1004 07:21:16.252095 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:16 crc kubenswrapper[4987]: E1004 07:21:16.277872 4987 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(2c460fd49fe066891319088bdec055d98d652bdaf8f532da50b7ae7ecaf161de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 07:21:16 crc kubenswrapper[4987]: E1004 07:21:16.277973 4987 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(2c460fd49fe066891319088bdec055d98d652bdaf8f532da50b7ae7ecaf161de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:16 crc kubenswrapper[4987]: E1004 07:21:16.278004 4987 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(2c460fd49fe066891319088bdec055d98d652bdaf8f532da50b7ae7ecaf161de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:16 crc kubenswrapper[4987]: E1004 07:21:16.278067 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace(92de2dd9-3957-4f74-a21e-ed5b040b5de4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace(92de2dd9-3957-4f74-a21e-ed5b040b5de4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(2c460fd49fe066891319088bdec055d98d652bdaf8f532da50b7ae7ecaf161de): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" Oct 04 07:21:21 crc kubenswrapper[4987]: I1004 07:21:21.053999 4987 scope.go:117] "RemoveContainer" containerID="1db87dc841e81f4e40e9a6c80e98c312d2b1ab807484d90d9f2e1be091bf01ad" Oct 04 07:21:21 crc kubenswrapper[4987]: E1004 07:21:21.054237 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-mc8jj_openshift-multus(659797bb-8ec1-4ad8-880e-7aec3b068c46)\"" pod="openshift-multus/multus-mc8jj" podUID="659797bb-8ec1-4ad8-880e-7aec3b068c46" Oct 04 07:21:28 crc kubenswrapper[4987]: I1004 07:21:28.053663 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:28 crc kubenswrapper[4987]: I1004 07:21:28.054770 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:28 crc kubenswrapper[4987]: E1004 07:21:28.087981 4987 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(e29469b064cba71ce0bf82fd6fe43c6ce2d15e2997204c2edba2ca9db4a8916a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 07:21:28 crc kubenswrapper[4987]: E1004 07:21:28.088087 4987 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(e29469b064cba71ce0bf82fd6fe43c6ce2d15e2997204c2edba2ca9db4a8916a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:28 crc kubenswrapper[4987]: E1004 07:21:28.088127 4987 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(e29469b064cba71ce0bf82fd6fe43c6ce2d15e2997204c2edba2ca9db4a8916a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:28 crc kubenswrapper[4987]: E1004 07:21:28.088212 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace(92de2dd9-3957-4f74-a21e-ed5b040b5de4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace(92de2dd9-3957-4f74-a21e-ed5b040b5de4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_openshift-marketplace_92de2dd9-3957-4f74-a21e-ed5b040b5de4_0(e29469b064cba71ce0bf82fd6fe43c6ce2d15e2997204c2edba2ca9db4a8916a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" Oct 04 07:21:34 crc kubenswrapper[4987]: I1004 07:21:34.056406 4987 scope.go:117] "RemoveContainer" containerID="1db87dc841e81f4e40e9a6c80e98c312d2b1ab807484d90d9f2e1be091bf01ad" Oct 04 07:21:34 crc kubenswrapper[4987]: I1004 07:21:34.350062 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-mc8jj_659797bb-8ec1-4ad8-880e-7aec3b068c46/kube-multus/2.log" Oct 04 07:21:34 crc kubenswrapper[4987]: I1004 07:21:34.350753 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-mc8jj" event={"ID":"659797bb-8ec1-4ad8-880e-7aec3b068c46","Type":"ContainerStarted","Data":"1e05331175365af08b91d45c628851e0e82cfc295fa075a88ba54b97d39e083d"} Oct 04 07:21:37 crc kubenswrapper[4987]: I1004 07:21:37.230393 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r2l62" Oct 04 07:21:40 crc kubenswrapper[4987]: I1004 07:21:40.053143 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:40 crc kubenswrapper[4987]: I1004 07:21:40.054041 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:40 crc kubenswrapper[4987]: I1004 07:21:40.265214 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn"] Oct 04 07:21:40 crc kubenswrapper[4987]: I1004 07:21:40.401457 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" event={"ID":"92de2dd9-3957-4f74-a21e-ed5b040b5de4","Type":"ContainerStarted","Data":"ca041136f37f0dafcae095a32165ef5599a1b83a21e5d546023bdaa2d409a31b"} Oct 04 07:21:41 crc kubenswrapper[4987]: I1004 07:21:41.412375 4987 generic.go:334] "Generic (PLEG): container finished" podID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerID="e3e5743b203bc7527ccc045217f45dccb2842ca0a0a28aa4d237b82097297ae4" exitCode=0 Oct 04 07:21:41 crc kubenswrapper[4987]: I1004 07:21:41.412435 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" event={"ID":"92de2dd9-3957-4f74-a21e-ed5b040b5de4","Type":"ContainerDied","Data":"e3e5743b203bc7527ccc045217f45dccb2842ca0a0a28aa4d237b82097297ae4"} Oct 04 07:21:41 crc kubenswrapper[4987]: I1004 07:21:41.415070 4987 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 07:21:42 crc kubenswrapper[4987]: I1004 07:21:42.703667 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:21:42 crc kubenswrapper[4987]: I1004 07:21:42.704287 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:21:43 crc kubenswrapper[4987]: I1004 07:21:43.426518 4987 generic.go:334] "Generic (PLEG): container finished" podID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerID="78159cf544b56aa7e8ad7d7b5c55af2d6f972e083298846008d83dc00c8e063b" exitCode=0 Oct 04 07:21:43 crc kubenswrapper[4987]: I1004 07:21:43.426639 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" event={"ID":"92de2dd9-3957-4f74-a21e-ed5b040b5de4","Type":"ContainerDied","Data":"78159cf544b56aa7e8ad7d7b5c55af2d6f972e083298846008d83dc00c8e063b"} Oct 04 07:21:44 crc kubenswrapper[4987]: I1004 07:21:44.432823 4987 generic.go:334] "Generic (PLEG): container finished" podID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerID="dbaf5c1a97935363de129840af03424d41fceea1e7b3c2973cba57b233941767" exitCode=0 Oct 04 07:21:44 crc kubenswrapper[4987]: I1004 07:21:44.432870 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" event={"ID":"92de2dd9-3957-4f74-a21e-ed5b040b5de4","Type":"ContainerDied","Data":"dbaf5c1a97935363de129840af03424d41fceea1e7b3c2973cba57b233941767"} Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.706060 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.854600 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-bundle\") pod \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.854681 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skps2\" (UniqueName: \"kubernetes.io/projected/92de2dd9-3957-4f74-a21e-ed5b040b5de4-kube-api-access-skps2\") pod \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.854711 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-util\") pod \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\" (UID: \"92de2dd9-3957-4f74-a21e-ed5b040b5de4\") " Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.856381 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-bundle" (OuterVolumeSpecName: "bundle") pod "92de2dd9-3957-4f74-a21e-ed5b040b5de4" (UID: "92de2dd9-3957-4f74-a21e-ed5b040b5de4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.861615 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92de2dd9-3957-4f74-a21e-ed5b040b5de4-kube-api-access-skps2" (OuterVolumeSpecName: "kube-api-access-skps2") pod "92de2dd9-3957-4f74-a21e-ed5b040b5de4" (UID: "92de2dd9-3957-4f74-a21e-ed5b040b5de4"). InnerVolumeSpecName "kube-api-access-skps2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.957244 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:45 crc kubenswrapper[4987]: I1004 07:21:45.957306 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skps2\" (UniqueName: \"kubernetes.io/projected/92de2dd9-3957-4f74-a21e-ed5b040b5de4-kube-api-access-skps2\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:46 crc kubenswrapper[4987]: I1004 07:21:46.244684 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-util" (OuterVolumeSpecName: "util") pod "92de2dd9-3957-4f74-a21e-ed5b040b5de4" (UID: "92de2dd9-3957-4f74-a21e-ed5b040b5de4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:21:46 crc kubenswrapper[4987]: I1004 07:21:46.261447 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/92de2dd9-3957-4f74-a21e-ed5b040b5de4-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:21:46 crc kubenswrapper[4987]: I1004 07:21:46.449404 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" event={"ID":"92de2dd9-3957-4f74-a21e-ed5b040b5de4","Type":"ContainerDied","Data":"ca041136f37f0dafcae095a32165ef5599a1b83a21e5d546023bdaa2d409a31b"} Oct 04 07:21:46 crc kubenswrapper[4987]: I1004 07:21:46.449481 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca041136f37f0dafcae095a32165ef5599a1b83a21e5d546023bdaa2d409a31b" Oct 04 07:21:46 crc kubenswrapper[4987]: I1004 07:21:46.449700 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.797080 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt"] Oct 04 07:21:58 crc kubenswrapper[4987]: E1004 07:21:58.797745 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="util" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.797758 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="util" Oct 04 07:21:58 crc kubenswrapper[4987]: E1004 07:21:58.797771 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="extract" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.797777 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="extract" Oct 04 07:21:58 crc kubenswrapper[4987]: E1004 07:21:58.797788 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="pull" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.797793 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="pull" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.797896 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="92de2dd9-3957-4f74-a21e-ed5b040b5de4" containerName="extract" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.798291 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.802011 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.802219 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.802301 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.802367 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-n27c6" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.802516 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.816861 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt"] Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.852904 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8hbz\" (UniqueName: \"kubernetes.io/projected/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-kube-api-access-r8hbz\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.852959 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-webhook-cert\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.852996 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-apiservice-cert\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.953972 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8hbz\" (UniqueName: \"kubernetes.io/projected/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-kube-api-access-r8hbz\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.954035 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-webhook-cert\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.954086 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-apiservice-cert\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.962679 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-webhook-cert\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.962708 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-apiservice-cert\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:58 crc kubenswrapper[4987]: I1004 07:21:58.979706 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8hbz\" (UniqueName: \"kubernetes.io/projected/dfbc115b-ff8f-424d-9f9d-d5e21ed2a790-kube-api-access-r8hbz\") pod \"metallb-operator-controller-manager-7f485cfd66-xwgrt\" (UID: \"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790\") " pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.114331 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.155416 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb"] Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.157095 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.163372 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-mwzk7" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.163571 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.163688 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.181474 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb"] Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.264372 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1222617f-db7d-4014-b952-79765d041405-apiservice-cert\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.264452 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1222617f-db7d-4014-b952-79765d041405-webhook-cert\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.264809 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tbf5\" (UniqueName: \"kubernetes.io/projected/1222617f-db7d-4014-b952-79765d041405-kube-api-access-2tbf5\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.367311 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1222617f-db7d-4014-b952-79765d041405-apiservice-cert\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.367424 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1222617f-db7d-4014-b952-79765d041405-webhook-cert\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.367448 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tbf5\" (UniqueName: \"kubernetes.io/projected/1222617f-db7d-4014-b952-79765d041405-kube-api-access-2tbf5\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.373489 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1222617f-db7d-4014-b952-79765d041405-webhook-cert\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.373497 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1222617f-db7d-4014-b952-79765d041405-apiservice-cert\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.397808 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tbf5\" (UniqueName: \"kubernetes.io/projected/1222617f-db7d-4014-b952-79765d041405-kube-api-access-2tbf5\") pod \"metallb-operator-webhook-server-6979fbb46b-pxmhb\" (UID: \"1222617f-db7d-4014-b952-79765d041405\") " pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.459590 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt"] Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.492704 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.539435 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" event={"ID":"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790","Type":"ContainerStarted","Data":"81c2ee473a5ec9b71bda7c8210b2787f80b41438290fd1368cce58ceb4e935e8"} Oct 04 07:21:59 crc kubenswrapper[4987]: I1004 07:21:59.683850 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb"] Oct 04 07:21:59 crc kubenswrapper[4987]: W1004 07:21:59.692967 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1222617f_db7d_4014_b952_79765d041405.slice/crio-0d4dc415d54034f5c84d68c2d6f7308edc67b10ef4f4aceb831ca78152ac381e WatchSource:0}: Error finding container 0d4dc415d54034f5c84d68c2d6f7308edc67b10ef4f4aceb831ca78152ac381e: Status 404 returned error can't find the container with id 0d4dc415d54034f5c84d68c2d6f7308edc67b10ef4f4aceb831ca78152ac381e Oct 04 07:22:00 crc kubenswrapper[4987]: I1004 07:22:00.547469 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" event={"ID":"1222617f-db7d-4014-b952-79765d041405","Type":"ContainerStarted","Data":"0d4dc415d54034f5c84d68c2d6f7308edc67b10ef4f4aceb831ca78152ac381e"} Oct 04 07:22:04 crc kubenswrapper[4987]: I1004 07:22:04.581860 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" event={"ID":"1222617f-db7d-4014-b952-79765d041405","Type":"ContainerStarted","Data":"d642a17ecc6e29ba5b336fca852c55507081002ee69e8c260244116cc40d0a12"} Oct 04 07:22:04 crc kubenswrapper[4987]: I1004 07:22:04.582440 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:22:04 crc kubenswrapper[4987]: I1004 07:22:04.583380 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" event={"ID":"dfbc115b-ff8f-424d-9f9d-d5e21ed2a790","Type":"ContainerStarted","Data":"fe83732eb486ca3e88b2d28a9e93f54c0d7d41bb42a65b3108f8c04a7d05f0ba"} Oct 04 07:22:04 crc kubenswrapper[4987]: I1004 07:22:04.583462 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:22:04 crc kubenswrapper[4987]: I1004 07:22:04.606867 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" podStartSLOduration=1.323385954 podStartE2EDuration="5.606848831s" podCreationTimestamp="2025-10-04 07:21:59 +0000 UTC" firstStartedPulling="2025-10-04 07:21:59.695901635 +0000 UTC m=+736.750800477" lastFinishedPulling="2025-10-04 07:22:03.979364522 +0000 UTC m=+741.034263354" observedRunningTime="2025-10-04 07:22:04.605033822 +0000 UTC m=+741.659932674" watchObservedRunningTime="2025-10-04 07:22:04.606848831 +0000 UTC m=+741.661747673" Oct 04 07:22:04 crc kubenswrapper[4987]: I1004 07:22:04.631133 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" podStartSLOduration=2.36069173 podStartE2EDuration="6.631114365s" podCreationTimestamp="2025-10-04 07:21:58 +0000 UTC" firstStartedPulling="2025-10-04 07:21:59.478779411 +0000 UTC m=+736.533678253" lastFinishedPulling="2025-10-04 07:22:03.749202046 +0000 UTC m=+740.804100888" observedRunningTime="2025-10-04 07:22:04.627120177 +0000 UTC m=+741.682019019" watchObservedRunningTime="2025-10-04 07:22:04.631114365 +0000 UTC m=+741.686013207" Oct 04 07:22:12 crc kubenswrapper[4987]: I1004 07:22:12.703119 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:22:12 crc kubenswrapper[4987]: I1004 07:22:12.703477 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:22:15 crc kubenswrapper[4987]: I1004 07:22:15.265812 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tfd59"] Oct 04 07:22:15 crc kubenswrapper[4987]: I1004 07:22:15.266783 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" podUID="1a5130dd-df25-4642-8180-637b14f8ee4a" containerName="controller-manager" containerID="cri-o://d7e183cdfc60afa6a7d71f1996684fbfcfd2abb560856e02700fd43e51897590" gracePeriod=30 Oct 04 07:22:15 crc kubenswrapper[4987]: I1004 07:22:15.298947 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9"] Oct 04 07:22:15 crc kubenswrapper[4987]: I1004 07:22:15.299282 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" podUID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" containerName="route-controller-manager" containerID="cri-o://30737a8c7d18f26335057b68a32aacf050f557157139f6735cbcd6f256a55bbc" gracePeriod=30 Oct 04 07:22:17 crc kubenswrapper[4987]: I1004 07:22:17.655090 4987 generic.go:334] "Generic (PLEG): container finished" podID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" containerID="30737a8c7d18f26335057b68a32aacf050f557157139f6735cbcd6f256a55bbc" exitCode=0 Oct 04 07:22:17 crc kubenswrapper[4987]: I1004 07:22:17.655204 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" event={"ID":"b3206d6a-5f6e-4f19-9dac-93b2689734dc","Type":"ContainerDied","Data":"30737a8c7d18f26335057b68a32aacf050f557157139f6735cbcd6f256a55bbc"} Oct 04 07:22:17 crc kubenswrapper[4987]: I1004 07:22:17.656564 4987 generic.go:334] "Generic (PLEG): container finished" podID="1a5130dd-df25-4642-8180-637b14f8ee4a" containerID="d7e183cdfc60afa6a7d71f1996684fbfcfd2abb560856e02700fd43e51897590" exitCode=0 Oct 04 07:22:17 crc kubenswrapper[4987]: I1004 07:22:17.656594 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" event={"ID":"1a5130dd-df25-4642-8180-637b14f8ee4a","Type":"ContainerDied","Data":"d7e183cdfc60afa6a7d71f1996684fbfcfd2abb560856e02700fd43e51897590"} Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.041087 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.072795 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z"] Oct 04 07:22:18 crc kubenswrapper[4987]: E1004 07:22:18.073083 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" containerName="route-controller-manager" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.073105 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" containerName="route-controller-manager" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.073242 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" containerName="route-controller-manager" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.073724 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.088426 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z"] Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.108739 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-client-ca\") pod \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.108951 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z42jj\" (UniqueName: \"kubernetes.io/projected/b3206d6a-5f6e-4f19-9dac-93b2689734dc-kube-api-access-z42jj\") pod \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.109020 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3206d6a-5f6e-4f19-9dac-93b2689734dc-serving-cert\") pod \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.109043 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-config\") pod \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\" (UID: \"b3206d6a-5f6e-4f19-9dac-93b2689734dc\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.109360 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27060516-199c-40a9-8040-af58b76e8bc0-client-ca\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.109448 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27060516-199c-40a9-8040-af58b76e8bc0-config\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.109473 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt8f2\" (UniqueName: \"kubernetes.io/projected/27060516-199c-40a9-8040-af58b76e8bc0-kube-api-access-kt8f2\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.109510 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27060516-199c-40a9-8040-af58b76e8bc0-serving-cert\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.110035 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-client-ca" (OuterVolumeSpecName: "client-ca") pod "b3206d6a-5f6e-4f19-9dac-93b2689734dc" (UID: "b3206d6a-5f6e-4f19-9dac-93b2689734dc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.110574 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-config" (OuterVolumeSpecName: "config") pod "b3206d6a-5f6e-4f19-9dac-93b2689734dc" (UID: "b3206d6a-5f6e-4f19-9dac-93b2689734dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.123985 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3206d6a-5f6e-4f19-9dac-93b2689734dc-kube-api-access-z42jj" (OuterVolumeSpecName: "kube-api-access-z42jj") pod "b3206d6a-5f6e-4f19-9dac-93b2689734dc" (UID: "b3206d6a-5f6e-4f19-9dac-93b2689734dc"). InnerVolumeSpecName "kube-api-access-z42jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.125267 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3206d6a-5f6e-4f19-9dac-93b2689734dc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b3206d6a-5f6e-4f19-9dac-93b2689734dc" (UID: "b3206d6a-5f6e-4f19-9dac-93b2689734dc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.209982 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27060516-199c-40a9-8040-af58b76e8bc0-client-ca\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210089 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27060516-199c-40a9-8040-af58b76e8bc0-config\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210113 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt8f2\" (UniqueName: \"kubernetes.io/projected/27060516-199c-40a9-8040-af58b76e8bc0-kube-api-access-kt8f2\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210151 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27060516-199c-40a9-8040-af58b76e8bc0-serving-cert\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210195 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z42jj\" (UniqueName: \"kubernetes.io/projected/b3206d6a-5f6e-4f19-9dac-93b2689734dc-kube-api-access-z42jj\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210289 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3206d6a-5f6e-4f19-9dac-93b2689734dc-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210440 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.210971 4987 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b3206d6a-5f6e-4f19-9dac-93b2689734dc-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.211284 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27060516-199c-40a9-8040-af58b76e8bc0-client-ca\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.211457 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27060516-199c-40a9-8040-af58b76e8bc0-config\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.217287 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27060516-199c-40a9-8040-af58b76e8bc0-serving-cert\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.238586 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt8f2\" (UniqueName: \"kubernetes.io/projected/27060516-199c-40a9-8040-af58b76e8bc0-kube-api-access-kt8f2\") pod \"route-controller-manager-8566b57bfb-9x58z\" (UID: \"27060516-199c-40a9-8040-af58b76e8bc0\") " pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.362445 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.404207 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.532425 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wllzd\" (UniqueName: \"kubernetes.io/projected/1a5130dd-df25-4642-8180-637b14f8ee4a-kube-api-access-wllzd\") pod \"1a5130dd-df25-4642-8180-637b14f8ee4a\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.532981 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-client-ca\") pod \"1a5130dd-df25-4642-8180-637b14f8ee4a\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.533024 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a5130dd-df25-4642-8180-637b14f8ee4a-serving-cert\") pod \"1a5130dd-df25-4642-8180-637b14f8ee4a\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.533068 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-proxy-ca-bundles\") pod \"1a5130dd-df25-4642-8180-637b14f8ee4a\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.533098 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-config\") pod \"1a5130dd-df25-4642-8180-637b14f8ee4a\" (UID: \"1a5130dd-df25-4642-8180-637b14f8ee4a\") " Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.534720 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-client-ca" (OuterVolumeSpecName: "client-ca") pod "1a5130dd-df25-4642-8180-637b14f8ee4a" (UID: "1a5130dd-df25-4642-8180-637b14f8ee4a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.536372 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1a5130dd-df25-4642-8180-637b14f8ee4a" (UID: "1a5130dd-df25-4642-8180-637b14f8ee4a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.537208 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-config" (OuterVolumeSpecName: "config") pod "1a5130dd-df25-4642-8180-637b14f8ee4a" (UID: "1a5130dd-df25-4642-8180-637b14f8ee4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.537290 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a5130dd-df25-4642-8180-637b14f8ee4a-kube-api-access-wllzd" (OuterVolumeSpecName: "kube-api-access-wllzd") pod "1a5130dd-df25-4642-8180-637b14f8ee4a" (UID: "1a5130dd-df25-4642-8180-637b14f8ee4a"). InnerVolumeSpecName "kube-api-access-wllzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.537511 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a5130dd-df25-4642-8180-637b14f8ee4a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1a5130dd-df25-4642-8180-637b14f8ee4a" (UID: "1a5130dd-df25-4642-8180-637b14f8ee4a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.634976 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wllzd\" (UniqueName: \"kubernetes.io/projected/1a5130dd-df25-4642-8180-637b14f8ee4a-kube-api-access-wllzd\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.635028 4987 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.635047 4987 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a5130dd-df25-4642-8180-637b14f8ee4a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.635060 4987 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.635074 4987 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a5130dd-df25-4642-8180-637b14f8ee4a-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.640161 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z"] Oct 04 07:22:18 crc kubenswrapper[4987]: W1004 07:22:18.653537 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27060516_199c_40a9_8040_af58b76e8bc0.slice/crio-68103edb2bc41681eb850308ae8cb7996489da1c3704f0973e3879b0ddfc5a9f WatchSource:0}: Error finding container 68103edb2bc41681eb850308ae8cb7996489da1c3704f0973e3879b0ddfc5a9f: Status 404 returned error can't find the container with id 68103edb2bc41681eb850308ae8cb7996489da1c3704f0973e3879b0ddfc5a9f Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.665790 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.665818 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tfd59" event={"ID":"1a5130dd-df25-4642-8180-637b14f8ee4a","Type":"ContainerDied","Data":"5197262856d44f062f99a35252a3c26217bd724053d9ff5f55ffceea0c36e70e"} Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.665925 4987 scope.go:117] "RemoveContainer" containerID="d7e183cdfc60afa6a7d71f1996684fbfcfd2abb560856e02700fd43e51897590" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.667582 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" event={"ID":"27060516-199c-40a9-8040-af58b76e8bc0","Type":"ContainerStarted","Data":"68103edb2bc41681eb850308ae8cb7996489da1c3704f0973e3879b0ddfc5a9f"} Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.670117 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" event={"ID":"b3206d6a-5f6e-4f19-9dac-93b2689734dc","Type":"ContainerDied","Data":"f0a98fae1a2c3c121c8a125215f9aaf52e80201f856366ac4b81e665c1dd2e0f"} Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.670215 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.683727 4987 scope.go:117] "RemoveContainer" containerID="30737a8c7d18f26335057b68a32aacf050f557157139f6735cbcd6f256a55bbc" Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.715334 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tfd59"] Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.719041 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tfd59"] Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.724988 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9"] Oct 04 07:22:18 crc kubenswrapper[4987]: I1004 07:22:18.728081 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cp8x9"] Oct 04 07:22:19 crc kubenswrapper[4987]: I1004 07:22:19.498048 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6979fbb46b-pxmhb" Oct 04 07:22:19 crc kubenswrapper[4987]: I1004 07:22:19.679976 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" event={"ID":"27060516-199c-40a9-8040-af58b76e8bc0","Type":"ContainerStarted","Data":"6782daec4f5700317c147320bafe2c1cd15d0bdf6cf1c0f1ba632934e3e0f522"} Oct 04 07:22:19 crc kubenswrapper[4987]: I1004 07:22:19.680318 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:19 crc kubenswrapper[4987]: I1004 07:22:19.686819 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" Oct 04 07:22:19 crc kubenswrapper[4987]: I1004 07:22:19.700834 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8566b57bfb-9x58z" podStartSLOduration=4.70080753 podStartE2EDuration="4.70080753s" podCreationTimestamp="2025-10-04 07:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:22:19.698736065 +0000 UTC m=+756.753634907" watchObservedRunningTime="2025-10-04 07:22:19.70080753 +0000 UTC m=+756.755706372" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.061001 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a5130dd-df25-4642-8180-637b14f8ee4a" path="/var/lib/kubelet/pods/1a5130dd-df25-4642-8180-637b14f8ee4a/volumes" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.061953 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3206d6a-5f6e-4f19-9dac-93b2689734dc" path="/var/lib/kubelet/pods/b3206d6a-5f6e-4f19-9dac-93b2689734dc/volumes" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.387919 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54c4d7b986-9swwr"] Oct 04 07:22:20 crc kubenswrapper[4987]: E1004 07:22:20.388256 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a5130dd-df25-4642-8180-637b14f8ee4a" containerName="controller-manager" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.388276 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a5130dd-df25-4642-8180-637b14f8ee4a" containerName="controller-manager" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.388397 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a5130dd-df25-4642-8180-637b14f8ee4a" containerName="controller-manager" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.388960 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.392508 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.392635 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.392808 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.394761 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.395034 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.397218 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.412003 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54c4d7b986-9swwr"] Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.430111 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.463604 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccprp\" (UniqueName: \"kubernetes.io/projected/6e76d938-aca1-4c2b-a392-c55188adc758-kube-api-access-ccprp\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.463697 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-client-ca\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.463901 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-proxy-ca-bundles\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.464050 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-config\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.464205 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e76d938-aca1-4c2b-a392-c55188adc758-serving-cert\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.565252 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-config\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.565345 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e76d938-aca1-4c2b-a392-c55188adc758-serving-cert\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.565380 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccprp\" (UniqueName: \"kubernetes.io/projected/6e76d938-aca1-4c2b-a392-c55188adc758-kube-api-access-ccprp\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.565419 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-client-ca\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.565452 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-proxy-ca-bundles\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.566876 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-proxy-ca-bundles\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.567005 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-config\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.567159 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e76d938-aca1-4c2b-a392-c55188adc758-client-ca\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.578364 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e76d938-aca1-4c2b-a392-c55188adc758-serving-cert\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.591079 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccprp\" (UniqueName: \"kubernetes.io/projected/6e76d938-aca1-4c2b-a392-c55188adc758-kube-api-access-ccprp\") pod \"controller-manager-54c4d7b986-9swwr\" (UID: \"6e76d938-aca1-4c2b-a392-c55188adc758\") " pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.705914 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:20 crc kubenswrapper[4987]: I1004 07:22:20.943542 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54c4d7b986-9swwr"] Oct 04 07:22:20 crc kubenswrapper[4987]: W1004 07:22:20.951346 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e76d938_aca1_4c2b_a392_c55188adc758.slice/crio-4050383b44de049fc02cdb10b04346cbff3bfee8b7b5a95b090506729e1553c8 WatchSource:0}: Error finding container 4050383b44de049fc02cdb10b04346cbff3bfee8b7b5a95b090506729e1553c8: Status 404 returned error can't find the container with id 4050383b44de049fc02cdb10b04346cbff3bfee8b7b5a95b090506729e1553c8 Oct 04 07:22:21 crc kubenswrapper[4987]: I1004 07:22:21.692515 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" event={"ID":"6e76d938-aca1-4c2b-a392-c55188adc758","Type":"ContainerStarted","Data":"487a6bff5c1d5b454d7eca5258d4694264dec20f6f6ad27f77deaa1e3f78acbd"} Oct 04 07:22:21 crc kubenswrapper[4987]: I1004 07:22:21.692930 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" event={"ID":"6e76d938-aca1-4c2b-a392-c55188adc758","Type":"ContainerStarted","Data":"4050383b44de049fc02cdb10b04346cbff3bfee8b7b5a95b090506729e1553c8"} Oct 04 07:22:22 crc kubenswrapper[4987]: I1004 07:22:22.698988 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:22 crc kubenswrapper[4987]: I1004 07:22:22.708516 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" Oct 04 07:22:22 crc kubenswrapper[4987]: I1004 07:22:22.730468 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54c4d7b986-9swwr" podStartSLOduration=7.730437538 podStartE2EDuration="7.730437538s" podCreationTimestamp="2025-10-04 07:22:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:22:21.714948835 +0000 UTC m=+758.769847667" watchObservedRunningTime="2025-10-04 07:22:22.730437538 +0000 UTC m=+759.785336380" Oct 04 07:22:23 crc kubenswrapper[4987]: I1004 07:22:23.663347 4987 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.118015 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7f485cfd66-xwgrt" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.874657 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh"] Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.875309 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.879639 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.879679 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-bmcgx" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.888165 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-fdd4g"] Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.912772 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh"] Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.912888 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.915446 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917580 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-reloader\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917672 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-conf\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917708 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-metrics-certs\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917741 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-startup\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917760 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-metrics\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917790 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpmhw\" (UniqueName: \"kubernetes.io/projected/e4a5ec9a-ead6-49d2-a31b-e28597d4c038-kube-api-access-rpmhw\") pod \"frr-k8s-webhook-server-64bf5d555-6lqfh\" (UID: \"e4a5ec9a-ead6-49d2-a31b-e28597d4c038\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917810 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9mjg\" (UniqueName: \"kubernetes.io/projected/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-kube-api-access-r9mjg\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917834 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e4a5ec9a-ead6-49d2-a31b-e28597d4c038-cert\") pod \"frr-k8s-webhook-server-64bf5d555-6lqfh\" (UID: \"e4a5ec9a-ead6-49d2-a31b-e28597d4c038\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.917855 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-sockets\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.920230 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.937498 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qj8gd"] Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.939831 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qj8gd" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.941711 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.943963 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-q2nw4" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.944020 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.943977 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.963948 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-kx8k7"] Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.965050 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.971364 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-kx8k7"] Oct 04 07:22:39 crc kubenswrapper[4987]: I1004 07:22:39.972537 4987 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.019930 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpmhw\" (UniqueName: \"kubernetes.io/projected/e4a5ec9a-ead6-49d2-a31b-e28597d4c038-kube-api-access-rpmhw\") pod \"frr-k8s-webhook-server-64bf5d555-6lqfh\" (UID: \"e4a5ec9a-ead6-49d2-a31b-e28597d4c038\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020011 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9mjg\" (UniqueName: \"kubernetes.io/projected/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-kube-api-access-r9mjg\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020059 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e4a5ec9a-ead6-49d2-a31b-e28597d4c038-cert\") pod \"frr-k8s-webhook-server-64bf5d555-6lqfh\" (UID: \"e4a5ec9a-ead6-49d2-a31b-e28597d4c038\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020091 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-sockets\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020111 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-reloader\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020140 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfgsv\" (UniqueName: \"kubernetes.io/projected/ebd2db86-e065-412a-aa7b-dd13f4359b1f-kube-api-access-mfgsv\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020168 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020202 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebd2db86-e065-412a-aa7b-dd13f4359b1f-metrics-certs\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020242 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebd2db86-e065-412a-aa7b-dd13f4359b1f-cert\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020277 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-conf\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020304 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-metrics-certs\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020340 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2grx\" (UniqueName: \"kubernetes.io/projected/47b35947-1821-4bf6-93cc-11c49e257911-kube-api-access-q2grx\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020363 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-startup\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020392 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-metrics\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020435 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-metrics-certs\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.020498 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/47b35947-1821-4bf6-93cc-11c49e257911-metallb-excludel2\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.021030 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-sockets\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.021213 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-conf\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.021891 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-frr-startup\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.024462 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-reloader\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.024325 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-metrics\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.029657 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-metrics-certs\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.051363 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e4a5ec9a-ead6-49d2-a31b-e28597d4c038-cert\") pod \"frr-k8s-webhook-server-64bf5d555-6lqfh\" (UID: \"e4a5ec9a-ead6-49d2-a31b-e28597d4c038\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.051553 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpmhw\" (UniqueName: \"kubernetes.io/projected/e4a5ec9a-ead6-49d2-a31b-e28597d4c038-kube-api-access-rpmhw\") pod \"frr-k8s-webhook-server-64bf5d555-6lqfh\" (UID: \"e4a5ec9a-ead6-49d2-a31b-e28597d4c038\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.055258 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9mjg\" (UniqueName: \"kubernetes.io/projected/e336de0c-93ab-4084-8d1a-1afcc48d7c0e-kube-api-access-r9mjg\") pod \"frr-k8s-fdd4g\" (UID: \"e336de0c-93ab-4084-8d1a-1afcc48d7c0e\") " pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.122849 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfgsv\" (UniqueName: \"kubernetes.io/projected/ebd2db86-e065-412a-aa7b-dd13f4359b1f-kube-api-access-mfgsv\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.122922 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.122947 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebd2db86-e065-412a-aa7b-dd13f4359b1f-metrics-certs\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.122976 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebd2db86-e065-412a-aa7b-dd13f4359b1f-cert\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.123039 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2grx\" (UniqueName: \"kubernetes.io/projected/47b35947-1821-4bf6-93cc-11c49e257911-kube-api-access-q2grx\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.123071 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-metrics-certs\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.123113 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/47b35947-1821-4bf6-93cc-11c49e257911-metallb-excludel2\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.124125 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/47b35947-1821-4bf6-93cc-11c49e257911-metallb-excludel2\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: E1004 07:22:40.124210 4987 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 04 07:22:40 crc kubenswrapper[4987]: E1004 07:22:40.124279 4987 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 07:22:40 crc kubenswrapper[4987]: E1004 07:22:40.124301 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-metrics-certs podName:47b35947-1821-4bf6-93cc-11c49e257911 nodeName:}" failed. No retries permitted until 2025-10-04 07:22:40.624280135 +0000 UTC m=+777.679178977 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-metrics-certs") pod "speaker-qj8gd" (UID: "47b35947-1821-4bf6-93cc-11c49e257911") : secret "speaker-certs-secret" not found Oct 04 07:22:40 crc kubenswrapper[4987]: E1004 07:22:40.124363 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist podName:47b35947-1821-4bf6-93cc-11c49e257911 nodeName:}" failed. No retries permitted until 2025-10-04 07:22:40.624334417 +0000 UTC m=+777.679233439 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist") pod "speaker-qj8gd" (UID: "47b35947-1821-4bf6-93cc-11c49e257911") : secret "metallb-memberlist" not found Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.128337 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebd2db86-e065-412a-aa7b-dd13f4359b1f-cert\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.128467 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebd2db86-e065-412a-aa7b-dd13f4359b1f-metrics-certs\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.143778 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfgsv\" (UniqueName: \"kubernetes.io/projected/ebd2db86-e065-412a-aa7b-dd13f4359b1f-kube-api-access-mfgsv\") pod \"controller-68d546b9d8-kx8k7\" (UID: \"ebd2db86-e065-412a-aa7b-dd13f4359b1f\") " pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.156338 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2grx\" (UniqueName: \"kubernetes.io/projected/47b35947-1821-4bf6-93cc-11c49e257911-kube-api-access-q2grx\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.215211 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.255119 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.291314 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.630926 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.631276 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-metrics-certs\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: E1004 07:22:40.631194 4987 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 07:22:40 crc kubenswrapper[4987]: E1004 07:22:40.631412 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist podName:47b35947-1821-4bf6-93cc-11c49e257911 nodeName:}" failed. No retries permitted until 2025-10-04 07:22:41.631385048 +0000 UTC m=+778.686283890 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist") pod "speaker-qj8gd" (UID: "47b35947-1821-4bf6-93cc-11c49e257911") : secret "metallb-memberlist" not found Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.636458 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-metrics-certs\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.737534 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh"] Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.794497 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-kx8k7"] Oct 04 07:22:40 crc kubenswrapper[4987]: W1004 07:22:40.800988 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebd2db86_e065_412a_aa7b_dd13f4359b1f.slice/crio-87aefa9e591e501b47be01149537a813abfaf4821d1662c358bcb13bc5e09a49 WatchSource:0}: Error finding container 87aefa9e591e501b47be01149537a813abfaf4821d1662c358bcb13bc5e09a49: Status 404 returned error can't find the container with id 87aefa9e591e501b47be01149537a813abfaf4821d1662c358bcb13bc5e09a49 Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.801573 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"172479ef0bf2ab91410fe2d02798f9746d3214bb510513aa0aee7395f97b59b6"} Oct 04 07:22:40 crc kubenswrapper[4987]: I1004 07:22:40.802819 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" event={"ID":"e4a5ec9a-ead6-49d2-a31b-e28597d4c038","Type":"ContainerStarted","Data":"d264944e6b18f759fa8ded710688151ea2be91097b84c91e2a3151c59e9e5b97"} Oct 04 07:22:41 crc kubenswrapper[4987]: I1004 07:22:41.644962 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:41 crc kubenswrapper[4987]: I1004 07:22:41.650884 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/47b35947-1821-4bf6-93cc-11c49e257911-memberlist\") pod \"speaker-qj8gd\" (UID: \"47b35947-1821-4bf6-93cc-11c49e257911\") " pod="metallb-system/speaker-qj8gd" Oct 04 07:22:41 crc kubenswrapper[4987]: I1004 07:22:41.765889 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qj8gd" Oct 04 07:22:41 crc kubenswrapper[4987]: W1004 07:22:41.785348 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47b35947_1821_4bf6_93cc_11c49e257911.slice/crio-7b06babfca96d6ff945cdb2eb045fd194b68b4e3409e735a055915b292121932 WatchSource:0}: Error finding container 7b06babfca96d6ff945cdb2eb045fd194b68b4e3409e735a055915b292121932: Status 404 returned error can't find the container with id 7b06babfca96d6ff945cdb2eb045fd194b68b4e3409e735a055915b292121932 Oct 04 07:22:41 crc kubenswrapper[4987]: I1004 07:22:41.810948 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qj8gd" event={"ID":"47b35947-1821-4bf6-93cc-11c49e257911","Type":"ContainerStarted","Data":"7b06babfca96d6ff945cdb2eb045fd194b68b4e3409e735a055915b292121932"} Oct 04 07:22:41 crc kubenswrapper[4987]: I1004 07:22:41.813005 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kx8k7" event={"ID":"ebd2db86-e065-412a-aa7b-dd13f4359b1f","Type":"ContainerStarted","Data":"78ec9a490adc5dc80c3383578862289fdb379c935e7179b18b881c7a87f1945b"} Oct 04 07:22:41 crc kubenswrapper[4987]: I1004 07:22:41.813085 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kx8k7" event={"ID":"ebd2db86-e065-412a-aa7b-dd13f4359b1f","Type":"ContainerStarted","Data":"87aefa9e591e501b47be01149537a813abfaf4821d1662c358bcb13bc5e09a49"} Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.161245 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lxzk2"] Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.163976 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.173545 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxzk2"] Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.268661 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-utilities\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.268734 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xntk2\" (UniqueName: \"kubernetes.io/projected/a801564a-9e19-465e-a51a-55c0837f581b-kube-api-access-xntk2\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.268773 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-catalog-content\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.369732 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-utilities\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.369820 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xntk2\" (UniqueName: \"kubernetes.io/projected/a801564a-9e19-465e-a51a-55c0837f581b-kube-api-access-xntk2\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.369852 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-catalog-content\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.370259 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-utilities\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.370435 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-catalog-content\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.406111 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xntk2\" (UniqueName: \"kubernetes.io/projected/a801564a-9e19-465e-a51a-55c0837f581b-kube-api-access-xntk2\") pod \"redhat-operators-lxzk2\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.499161 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.704990 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.705333 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.705385 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.706069 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"241b883c886b42b50a0701b97bca81215e874c5bdbc8a5d84d217f37244c27ff"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.706130 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://241b883c886b42b50a0701b97bca81215e874c5bdbc8a5d84d217f37244c27ff" gracePeriod=600 Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.853519 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qj8gd" event={"ID":"47b35947-1821-4bf6-93cc-11c49e257911","Type":"ContainerStarted","Data":"8789fb67fe7a8971a910c685ecff8c0e12e69251577d1b7607aed6f5ee85ef58"} Oct 04 07:22:42 crc kubenswrapper[4987]: I1004 07:22:42.892444 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lxzk2"] Oct 04 07:22:43 crc kubenswrapper[4987]: I1004 07:22:43.863972 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="241b883c886b42b50a0701b97bca81215e874c5bdbc8a5d84d217f37244c27ff" exitCode=0 Oct 04 07:22:43 crc kubenswrapper[4987]: I1004 07:22:43.864193 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"241b883c886b42b50a0701b97bca81215e874c5bdbc8a5d84d217f37244c27ff"} Oct 04 07:22:43 crc kubenswrapper[4987]: I1004 07:22:43.864436 4987 scope.go:117] "RemoveContainer" containerID="505ec6340e253f2b31b71cdfb3e0b7658643e55c35199ae6c3e39c731aca0c48" Oct 04 07:22:43 crc kubenswrapper[4987]: I1004 07:22:43.875971 4987 generic.go:334] "Generic (PLEG): container finished" podID="a801564a-9e19-465e-a51a-55c0837f581b" containerID="e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2" exitCode=0 Oct 04 07:22:43 crc kubenswrapper[4987]: I1004 07:22:43.876039 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerDied","Data":"e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2"} Oct 04 07:22:43 crc kubenswrapper[4987]: I1004 07:22:43.876103 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerStarted","Data":"7a2d0385eb55c457a43995ec51e2716f85816e82b91e00964237d67369bd4b44"} Oct 04 07:22:44 crc kubenswrapper[4987]: I1004 07:22:44.886837 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"914ae0228b00a348da279a19d0c789018ccba172664058b29703314fa8fd5ab8"} Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.658617 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v9n7c"] Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.661519 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.668820 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9n7c"] Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.859467 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hq5h\" (UniqueName: \"kubernetes.io/projected/281f644d-ab95-4743-918b-660dd2644aa6-kube-api-access-5hq5h\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.859686 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281f644d-ab95-4743-918b-660dd2644aa6-utilities\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.859908 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281f644d-ab95-4743-918b-660dd2644aa6-catalog-content\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.961023 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281f644d-ab95-4743-918b-660dd2644aa6-catalog-content\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.961149 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hq5h\" (UniqueName: \"kubernetes.io/projected/281f644d-ab95-4743-918b-660dd2644aa6-kube-api-access-5hq5h\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.961204 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281f644d-ab95-4743-918b-660dd2644aa6-utilities\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.961733 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/281f644d-ab95-4743-918b-660dd2644aa6-catalog-content\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.961957 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/281f644d-ab95-4743-918b-660dd2644aa6-utilities\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:52 crc kubenswrapper[4987]: I1004 07:22:52.999740 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hq5h\" (UniqueName: \"kubernetes.io/projected/281f644d-ab95-4743-918b-660dd2644aa6-kube-api-access-5hq5h\") pod \"community-operators-v9n7c\" (UID: \"281f644d-ab95-4743-918b-660dd2644aa6\") " pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:22:53 crc kubenswrapper[4987]: I1004 07:22:53.289360 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:23:00 crc kubenswrapper[4987]: E1004 07:23:00.414174 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:41205f57dd50b222640776ca5fcda336ca1541f53dae820d7bc6669f52c28a87" Oct 04 07:23:00 crc kubenswrapper[4987]: E1004 07:23:00.414658 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:cp-frr-files,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:41205f57dd50b222640776ca5fcda336ca1541f53dae820d7bc6669f52c28a87,Command:[/bin/sh -c cp -rLf /tmp/frr/* /etc/frr/],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:frr-startup,ReadOnly:false,MountPath:/tmp/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:frr-conf,ReadOnly:false,MountPath:/etc/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r9mjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*100,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*101,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-fdd4g_metallb-system(e336de0c-93ab-4084-8d1a-1afcc48d7c0e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:23:00 crc kubenswrapper[4987]: E1004 07:23:00.416306 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-fdd4g" podUID="e336de0c-93ab-4084-8d1a-1afcc48d7c0e" Oct 04 07:23:00 crc kubenswrapper[4987]: E1004 07:23:00.427312 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:41205f57dd50b222640776ca5fcda336ca1541f53dae820d7bc6669f52c28a87" Oct 04 07:23:00 crc kubenswrapper[4987]: E1004 07:23:00.427684 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:frr-k8s-webhook-server,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:41205f57dd50b222640776ca5fcda336ca1541f53dae820d7bc6669f52c28a87,Command:[/frr-k8s],Args:[--log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=$(NAMESPACE) --metrics-bind-address=:7572],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:monitoring,HostPort:0,ContainerPort:7572,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rpmhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000680000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-webhook-server-64bf5d555-6lqfh_metallb-system(e4a5ec9a-ead6-49d2-a31b-e28597d4c038): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 07:23:00 crc kubenswrapper[4987]: E1004 07:23:00.428903 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" podUID="e4a5ec9a-ead6-49d2-a31b-e28597d4c038" Oct 04 07:23:01 crc kubenswrapper[4987]: E1004 07:23:01.933114 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:41205f57dd50b222640776ca5fcda336ca1541f53dae820d7bc6669f52c28a87\\\"\"" pod="metallb-system/frr-k8s-fdd4g" podUID="e336de0c-93ab-4084-8d1a-1afcc48d7c0e" Oct 04 07:23:01 crc kubenswrapper[4987]: E1004 07:23:01.934271 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:41205f57dd50b222640776ca5fcda336ca1541f53dae820d7bc6669f52c28a87\\\"\"" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" podUID="e4a5ec9a-ead6-49d2-a31b-e28597d4c038" Oct 04 07:23:02 crc kubenswrapper[4987]: I1004 07:23:02.865217 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9n7c"] Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.000378 4987 generic.go:334] "Generic (PLEG): container finished" podID="281f644d-ab95-4743-918b-660dd2644aa6" containerID="e9ad2810b6a83197202057bcc353ecf8b0651505d48e878f0f7282ff36409d6b" exitCode=0 Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.000457 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9n7c" event={"ID":"281f644d-ab95-4743-918b-660dd2644aa6","Type":"ContainerDied","Data":"e9ad2810b6a83197202057bcc353ecf8b0651505d48e878f0f7282ff36409d6b"} Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.000877 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9n7c" event={"ID":"281f644d-ab95-4743-918b-660dd2644aa6","Type":"ContainerStarted","Data":"b3dd96b228b7a4cb4b787c848c40a0a7c2772548ffcdcb33aa3ad01690093faf"} Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.003340 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qj8gd" event={"ID":"47b35947-1821-4bf6-93cc-11c49e257911","Type":"ContainerStarted","Data":"be6fb32aebb28faf20bbfe12ff7e4cefdbffa6eef1df1f40c351db066f81f1fb"} Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.003479 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qj8gd" Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.006647 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-kx8k7" event={"ID":"ebd2db86-e065-412a-aa7b-dd13f4359b1f","Type":"ContainerStarted","Data":"dc1e6278908b7e139b8573be88d241b23b05df09d3f22bffcb55829e811abb14"} Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.006988 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.007862 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qj8gd" Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.008665 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerStarted","Data":"894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3"} Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.011906 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-kx8k7" Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.041612 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qj8gd" podStartSLOduration=3.894525788 podStartE2EDuration="25.041582726s" podCreationTimestamp="2025-10-04 07:22:39 +0000 UTC" firstStartedPulling="2025-10-04 07:22:42.088050289 +0000 UTC m=+779.142949131" lastFinishedPulling="2025-10-04 07:23:03.235107217 +0000 UTC m=+800.290006069" observedRunningTime="2025-10-04 07:23:04.037210878 +0000 UTC m=+801.092109730" watchObservedRunningTime="2025-10-04 07:23:04.041582726 +0000 UTC m=+801.096481568" Oct 04 07:23:04 crc kubenswrapper[4987]: I1004 07:23:04.062472 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-kx8k7" podStartSLOduration=2.811630012 podStartE2EDuration="25.062456018s" podCreationTimestamp="2025-10-04 07:22:39 +0000 UTC" firstStartedPulling="2025-10-04 07:22:40.985230591 +0000 UTC m=+778.040129433" lastFinishedPulling="2025-10-04 07:23:03.236056597 +0000 UTC m=+800.290955439" observedRunningTime="2025-10-04 07:23:04.061081174 +0000 UTC m=+801.115980016" watchObservedRunningTime="2025-10-04 07:23:04.062456018 +0000 UTC m=+801.117354860" Oct 04 07:23:05 crc kubenswrapper[4987]: I1004 07:23:05.017396 4987 generic.go:334] "Generic (PLEG): container finished" podID="a801564a-9e19-465e-a51a-55c0837f581b" containerID="894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3" exitCode=0 Oct 04 07:23:05 crc kubenswrapper[4987]: I1004 07:23:05.017479 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerDied","Data":"894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3"} Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.058718 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerStarted","Data":"74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379"} Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.075902 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lxzk2" podStartSLOduration=3.234990363 podStartE2EDuration="29.075885484s" podCreationTimestamp="2025-10-04 07:22:42 +0000 UTC" firstStartedPulling="2025-10-04 07:22:43.877813797 +0000 UTC m=+780.932712639" lastFinishedPulling="2025-10-04 07:23:09.718708918 +0000 UTC m=+806.773607760" observedRunningTime="2025-10-04 07:23:11.075545433 +0000 UTC m=+808.130444285" watchObservedRunningTime="2025-10-04 07:23:11.075885484 +0000 UTC m=+808.130784326" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.401211 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-5mp8v"] Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.402121 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.405316 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-x58hw" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.406678 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.408139 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.413536 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-5mp8v"] Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.449572 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9hbl\" (UniqueName: \"kubernetes.io/projected/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0-kube-api-access-v9hbl\") pod \"mariadb-operator-index-5mp8v\" (UID: \"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0\") " pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.550555 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9hbl\" (UniqueName: \"kubernetes.io/projected/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0-kube-api-access-v9hbl\") pod \"mariadb-operator-index-5mp8v\" (UID: \"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0\") " pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.573437 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9hbl\" (UniqueName: \"kubernetes.io/projected/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0-kube-api-access-v9hbl\") pod \"mariadb-operator-index-5mp8v\" (UID: \"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0\") " pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:11 crc kubenswrapper[4987]: I1004 07:23:11.720802 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:12 crc kubenswrapper[4987]: I1004 07:23:12.500303 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:23:12 crc kubenswrapper[4987]: I1004 07:23:12.500795 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:23:12 crc kubenswrapper[4987]: I1004 07:23:12.746856 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-5mp8v"] Oct 04 07:23:12 crc kubenswrapper[4987]: W1004 07:23:12.764030 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8ecfc35_b0cc_4f19_8fc8_23e659b853e0.slice/crio-f2310800a2110a48ff1f601c14965a42b9009d7abfec11ecb2efdf8d60e50c8b WatchSource:0}: Error finding container f2310800a2110a48ff1f601c14965a42b9009d7abfec11ecb2efdf8d60e50c8b: Status 404 returned error can't find the container with id f2310800a2110a48ff1f601c14965a42b9009d7abfec11ecb2efdf8d60e50c8b Oct 04 07:23:13 crc kubenswrapper[4987]: I1004 07:23:13.071404 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5mp8v" event={"ID":"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0","Type":"ContainerStarted","Data":"f2310800a2110a48ff1f601c14965a42b9009d7abfec11ecb2efdf8d60e50c8b"} Oct 04 07:23:13 crc kubenswrapper[4987]: I1004 07:23:13.073138 4987 generic.go:334] "Generic (PLEG): container finished" podID="281f644d-ab95-4743-918b-660dd2644aa6" containerID="98e5ef06628867ec176f91bffaa3d4f89051f9db58b2928535d4911ff1d67a14" exitCode=0 Oct 04 07:23:13 crc kubenswrapper[4987]: I1004 07:23:13.074090 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9n7c" event={"ID":"281f644d-ab95-4743-918b-660dd2644aa6","Type":"ContainerDied","Data":"98e5ef06628867ec176f91bffaa3d4f89051f9db58b2928535d4911ff1d67a14"} Oct 04 07:23:13 crc kubenswrapper[4987]: I1004 07:23:13.537511 4987 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lxzk2" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="registry-server" probeResult="failure" output=< Oct 04 07:23:13 crc kubenswrapper[4987]: timeout: failed to connect service ":50051" within 1s Oct 04 07:23:13 crc kubenswrapper[4987]: > Oct 04 07:23:15 crc kubenswrapper[4987]: I1004 07:23:15.085144 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v9n7c" event={"ID":"281f644d-ab95-4743-918b-660dd2644aa6","Type":"ContainerStarted","Data":"d9ab04a5595d7b035526d26791aa7d78dae1b504638e4d29f903a73d8e530b0c"} Oct 04 07:23:16 crc kubenswrapper[4987]: I1004 07:23:16.888794 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-5mp8v"] Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.118072 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v9n7c" podStartSLOduration=14.449104735 podStartE2EDuration="25.118051435s" podCreationTimestamp="2025-10-04 07:22:52 +0000 UTC" firstStartedPulling="2025-10-04 07:23:04.002282029 +0000 UTC m=+801.057180871" lastFinishedPulling="2025-10-04 07:23:14.671228739 +0000 UTC m=+811.726127571" observedRunningTime="2025-10-04 07:23:17.113821091 +0000 UTC m=+814.168719933" watchObservedRunningTime="2025-10-04 07:23:17.118051435 +0000 UTC m=+814.172950277" Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.696926 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-n2x7n"] Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.697682 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.706123 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-n2x7n"] Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.729811 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj65k\" (UniqueName: \"kubernetes.io/projected/092dd76b-510c-49fb-be46-5a8ebfb818a7-kube-api-access-cj65k\") pod \"mariadb-operator-index-n2x7n\" (UID: \"092dd76b-510c-49fb-be46-5a8ebfb818a7\") " pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.830864 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj65k\" (UniqueName: \"kubernetes.io/projected/092dd76b-510c-49fb-be46-5a8ebfb818a7-kube-api-access-cj65k\") pod \"mariadb-operator-index-n2x7n\" (UID: \"092dd76b-510c-49fb-be46-5a8ebfb818a7\") " pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:17 crc kubenswrapper[4987]: I1004 07:23:17.849650 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj65k\" (UniqueName: \"kubernetes.io/projected/092dd76b-510c-49fb-be46-5a8ebfb818a7-kube-api-access-cj65k\") pod \"mariadb-operator-index-n2x7n\" (UID: \"092dd76b-510c-49fb-be46-5a8ebfb818a7\") " pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:18 crc kubenswrapper[4987]: I1004 07:23:18.014457 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:18 crc kubenswrapper[4987]: I1004 07:23:18.406580 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-n2x7n"] Oct 04 07:23:19 crc kubenswrapper[4987]: I1004 07:23:19.112075 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-n2x7n" event={"ID":"092dd76b-510c-49fb-be46-5a8ebfb818a7","Type":"ContainerStarted","Data":"450a4b44780e7802e55e18ae167bf3252232476cdf1c981211d0e64ac5e284dd"} Oct 04 07:23:22 crc kubenswrapper[4987]: I1004 07:23:22.538639 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:23:22 crc kubenswrapper[4987]: I1004 07:23:22.579036 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.112768 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-67696"] Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.123751 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.141311 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-67696"] Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.290243 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.290303 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.307733 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk5jq\" (UniqueName: \"kubernetes.io/projected/8e0fb420-6ff0-412b-9a82-149ea4d941fb-kube-api-access-rk5jq\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.307801 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-catalog-content\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.307893 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-utilities\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.340905 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.409669 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk5jq\" (UniqueName: \"kubernetes.io/projected/8e0fb420-6ff0-412b-9a82-149ea4d941fb-kube-api-access-rk5jq\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.409739 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-catalog-content\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.409782 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-utilities\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.410511 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-catalog-content\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.410577 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-utilities\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.441375 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk5jq\" (UniqueName: \"kubernetes.io/projected/8e0fb420-6ff0-412b-9a82-149ea4d941fb-kube-api-access-rk5jq\") pod \"certified-operators-67696\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:23 crc kubenswrapper[4987]: I1004 07:23:23.454844 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:24 crc kubenswrapper[4987]: I1004 07:23:24.190762 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v9n7c" Oct 04 07:23:25 crc kubenswrapper[4987]: I1004 07:23:25.406311 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-67696"] Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.167459 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" event={"ID":"e4a5ec9a-ead6-49d2-a31b-e28597d4c038","Type":"ContainerStarted","Data":"52358aae0d801bb4c000d50289dd449161f14673ceda71b2eaa12ec54d4d479e"} Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.168225 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.170375 4987 generic.go:334] "Generic (PLEG): container finished" podID="e336de0c-93ab-4084-8d1a-1afcc48d7c0e" containerID="0a0ece1591fd32180f91561f8201cfba595c3d0ad24c637ef103a41e432f0996" exitCode=0 Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.170459 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerDied","Data":"0a0ece1591fd32180f91561f8201cfba595c3d0ad24c637ef103a41e432f0996"} Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.184392 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-n2x7n" event={"ID":"092dd76b-510c-49fb-be46-5a8ebfb818a7","Type":"ContainerStarted","Data":"3ec23bcde78a9fd6c5af1e8c6c92c18b944d5457ba8e6bf5d8dd5dc64f42d697"} Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.187246 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5mp8v" event={"ID":"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0","Type":"ContainerStarted","Data":"38f0e1f4cf99bf7e94ba2c554f1a28c95e6f15d93f4616bb8baed25168b80e8a"} Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.187275 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-5mp8v" podUID="b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" containerName="registry-server" containerID="cri-o://38f0e1f4cf99bf7e94ba2c554f1a28c95e6f15d93f4616bb8baed25168b80e8a" gracePeriod=2 Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.195909 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" podStartSLOduration=2.9209558490000003 podStartE2EDuration="47.195837368s" podCreationTimestamp="2025-10-04 07:22:39 +0000 UTC" firstStartedPulling="2025-10-04 07:22:40.750720133 +0000 UTC m=+777.805618975" lastFinishedPulling="2025-10-04 07:23:25.025601652 +0000 UTC m=+822.080500494" observedRunningTime="2025-10-04 07:23:26.193131612 +0000 UTC m=+823.248030474" watchObservedRunningTime="2025-10-04 07:23:26.195837368 +0000 UTC m=+823.250736210" Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.198076 4987 generic.go:334] "Generic (PLEG): container finished" podID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerID="d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264" exitCode=0 Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.198156 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67696" event={"ID":"8e0fb420-6ff0-412b-9a82-149ea4d941fb","Type":"ContainerDied","Data":"d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264"} Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.198199 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67696" event={"ID":"8e0fb420-6ff0-412b-9a82-149ea4d941fb","Type":"ContainerStarted","Data":"c7259ce2e9c4125442cf50e04b17e35e2b1f9deebd48ad3716bb51498863efc9"} Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.222097 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-5mp8v" podStartSLOduration=3.000484035 podStartE2EDuration="15.222058991s" podCreationTimestamp="2025-10-04 07:23:11 +0000 UTC" firstStartedPulling="2025-10-04 07:23:12.76633776 +0000 UTC m=+809.821236602" lastFinishedPulling="2025-10-04 07:23:24.987912716 +0000 UTC m=+822.042811558" observedRunningTime="2025-10-04 07:23:26.211590658 +0000 UTC m=+823.266489510" watchObservedRunningTime="2025-10-04 07:23:26.222058991 +0000 UTC m=+823.276957853" Oct 04 07:23:26 crc kubenswrapper[4987]: I1004 07:23:26.287771 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-n2x7n" podStartSLOduration=2.696470037 podStartE2EDuration="9.287752174s" podCreationTimestamp="2025-10-04 07:23:17 +0000 UTC" firstStartedPulling="2025-10-04 07:23:18.41621917 +0000 UTC m=+815.471118012" lastFinishedPulling="2025-10-04 07:23:25.007501307 +0000 UTC m=+822.062400149" observedRunningTime="2025-10-04 07:23:26.265841608 +0000 UTC m=+823.320740450" watchObservedRunningTime="2025-10-04 07:23:26.287752174 +0000 UTC m=+823.342651016" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.096506 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxzk2"] Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.097875 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lxzk2" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="registry-server" containerID="cri-o://74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379" gracePeriod=2 Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.206532 4987 generic.go:334] "Generic (PLEG): container finished" podID="b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" containerID="38f0e1f4cf99bf7e94ba2c554f1a28c95e6f15d93f4616bb8baed25168b80e8a" exitCode=0 Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.206701 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5mp8v" event={"ID":"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0","Type":"ContainerDied","Data":"38f0e1f4cf99bf7e94ba2c554f1a28c95e6f15d93f4616bb8baed25168b80e8a"} Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.210049 4987 generic.go:334] "Generic (PLEG): container finished" podID="e336de0c-93ab-4084-8d1a-1afcc48d7c0e" containerID="283029aa88b99be6407ac48fbf028dde5cd5ecda3e31689f4c7cab3cd13c7b17" exitCode=0 Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.210467 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerDied","Data":"283029aa88b99be6407ac48fbf028dde5cd5ecda3e31689f4c7cab3cd13c7b17"} Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.587141 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.684501 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xntk2\" (UniqueName: \"kubernetes.io/projected/a801564a-9e19-465e-a51a-55c0837f581b-kube-api-access-xntk2\") pod \"a801564a-9e19-465e-a51a-55c0837f581b\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.684596 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-utilities\") pod \"a801564a-9e19-465e-a51a-55c0837f581b\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.684704 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-catalog-content\") pod \"a801564a-9e19-465e-a51a-55c0837f581b\" (UID: \"a801564a-9e19-465e-a51a-55c0837f581b\") " Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.685760 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-utilities" (OuterVolumeSpecName: "utilities") pod "a801564a-9e19-465e-a51a-55c0837f581b" (UID: "a801564a-9e19-465e-a51a-55c0837f581b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.695218 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a801564a-9e19-465e-a51a-55c0837f581b-kube-api-access-xntk2" (OuterVolumeSpecName: "kube-api-access-xntk2") pod "a801564a-9e19-465e-a51a-55c0837f581b" (UID: "a801564a-9e19-465e-a51a-55c0837f581b"). InnerVolumeSpecName "kube-api-access-xntk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.727938 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.786780 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xntk2\" (UniqueName: \"kubernetes.io/projected/a801564a-9e19-465e-a51a-55c0837f581b-kube-api-access-xntk2\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.786828 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.788994 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a801564a-9e19-465e-a51a-55c0837f581b" (UID: "a801564a-9e19-465e-a51a-55c0837f581b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.887610 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9hbl\" (UniqueName: \"kubernetes.io/projected/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0-kube-api-access-v9hbl\") pod \"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0\" (UID: \"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0\") " Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.888024 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a801564a-9e19-465e-a51a-55c0837f581b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.890862 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0-kube-api-access-v9hbl" (OuterVolumeSpecName: "kube-api-access-v9hbl") pod "b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" (UID: "b8ecfc35-b0cc-4f19-8fc8-23e659b853e0"). InnerVolumeSpecName "kube-api-access-v9hbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:23:27 crc kubenswrapper[4987]: I1004 07:23:27.989793 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9hbl\" (UniqueName: \"kubernetes.io/projected/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0-kube-api-access-v9hbl\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.015406 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.015485 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.051225 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.217428 4987 generic.go:334] "Generic (PLEG): container finished" podID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerID="7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3" exitCode=0 Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.217498 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67696" event={"ID":"8e0fb420-6ff0-412b-9a82-149ea4d941fb","Type":"ContainerDied","Data":"7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3"} Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.219421 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-5mp8v" event={"ID":"b8ecfc35-b0cc-4f19-8fc8-23e659b853e0","Type":"ContainerDied","Data":"f2310800a2110a48ff1f601c14965a42b9009d7abfec11ecb2efdf8d60e50c8b"} Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.219473 4987 scope.go:117] "RemoveContainer" containerID="38f0e1f4cf99bf7e94ba2c554f1a28c95e6f15d93f4616bb8baed25168b80e8a" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.219574 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-5mp8v" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.229979 4987 generic.go:334] "Generic (PLEG): container finished" podID="a801564a-9e19-465e-a51a-55c0837f581b" containerID="74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379" exitCode=0 Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.230112 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lxzk2" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.230696 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerDied","Data":"74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379"} Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.230728 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lxzk2" event={"ID":"a801564a-9e19-465e-a51a-55c0837f581b","Type":"ContainerDied","Data":"7a2d0385eb55c457a43995ec51e2716f85816e82b91e00964237d67369bd4b44"} Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.237072 4987 generic.go:334] "Generic (PLEG): container finished" podID="e336de0c-93ab-4084-8d1a-1afcc48d7c0e" containerID="cb1045c0c2e661d4c508ef51e8a3eee67a8e79e8bf7b27192dcf029d930d2627" exitCode=0 Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.237133 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerDied","Data":"cb1045c0c2e661d4c508ef51e8a3eee67a8e79e8bf7b27192dcf029d930d2627"} Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.254765 4987 scope.go:117] "RemoveContainer" containerID="74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.275991 4987 scope.go:117] "RemoveContainer" containerID="894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.315731 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-5mp8v"] Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.320231 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-5mp8v"] Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.320636 4987 scope.go:117] "RemoveContainer" containerID="e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.327665 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lxzk2"] Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.335060 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lxzk2"] Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.351634 4987 scope.go:117] "RemoveContainer" containerID="74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379" Oct 04 07:23:28 crc kubenswrapper[4987]: E1004 07:23:28.352432 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379\": container with ID starting with 74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379 not found: ID does not exist" containerID="74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.352472 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379"} err="failed to get container status \"74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379\": rpc error: code = NotFound desc = could not find container \"74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379\": container with ID starting with 74bbfa16acd27684f1a80033bcadc350973bac3692942cddafc5686034ba5379 not found: ID does not exist" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.352503 4987 scope.go:117] "RemoveContainer" containerID="894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3" Oct 04 07:23:28 crc kubenswrapper[4987]: E1004 07:23:28.353096 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3\": container with ID starting with 894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3 not found: ID does not exist" containerID="894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.353134 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3"} err="failed to get container status \"894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3\": rpc error: code = NotFound desc = could not find container \"894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3\": container with ID starting with 894df6614adea6eaa6b793a861f0b5b076fa8ec7949d73b257d5f42e2646c7d3 not found: ID does not exist" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.353165 4987 scope.go:117] "RemoveContainer" containerID="e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2" Oct 04 07:23:28 crc kubenswrapper[4987]: E1004 07:23:28.353754 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2\": container with ID starting with e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2 not found: ID does not exist" containerID="e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2" Oct 04 07:23:28 crc kubenswrapper[4987]: I1004 07:23:28.353781 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2"} err="failed to get container status \"e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2\": rpc error: code = NotFound desc = could not find container \"e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2\": container with ID starting with e4141031ee23bc3e8c62eebd020f13b949ccaa7918703f9770912a6429f9c0b2 not found: ID does not exist" Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.251728 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67696" event={"ID":"8e0fb420-6ff0-412b-9a82-149ea4d941fb","Type":"ContainerStarted","Data":"c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e"} Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.258441 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"c61871db2e5e17ef012acd1d99e245d57b71758ff038b3ad52435512b06d1d00"} Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.258487 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"fe546d4d2677f4a41bb91b68f9bed37e91d2ede1594a3209bdb5c4c32a889f97"} Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.258497 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"489bbb9212cc8e44df9ed0dec8cad91f667fc54372026824a2117a7e096f076e"} Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.258505 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"7d6728ff15c344ff7e4949f9c8ad985c3e8867cf9ec54376780282aebbfa5083"} Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.258513 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"87bc7df0f147188e52b4a40f033f5c8f449026ab6ee83e0a43c447c988a4b0b0"} Oct 04 07:23:29 crc kubenswrapper[4987]: I1004 07:23:29.275328 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-67696" podStartSLOduration=3.716391718 podStartE2EDuration="6.27530837s" podCreationTimestamp="2025-10-04 07:23:23 +0000 UTC" firstStartedPulling="2025-10-04 07:23:26.201413685 +0000 UTC m=+823.256312537" lastFinishedPulling="2025-10-04 07:23:28.760330347 +0000 UTC m=+825.815229189" observedRunningTime="2025-10-04 07:23:29.271111348 +0000 UTC m=+826.326010210" watchObservedRunningTime="2025-10-04 07:23:29.27530837 +0000 UTC m=+826.330207212" Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.067069 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a801564a-9e19-465e-a51a-55c0837f581b" path="/var/lib/kubelet/pods/a801564a-9e19-465e-a51a-55c0837f581b/volumes" Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.067754 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" path="/var/lib/kubelet/pods/b8ecfc35-b0cc-4f19-8fc8-23e659b853e0/volumes" Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.113985 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v9n7c"] Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.268140 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-fdd4g" event={"ID":"e336de0c-93ab-4084-8d1a-1afcc48d7c0e","Type":"ContainerStarted","Data":"5da61651afd940febea515f6c26526bcf30a455df3f749839783a4a43f7c4763"} Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.686730 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-fdd4g" podStartSLOduration=7.120500618 podStartE2EDuration="51.686708146s" podCreationTimestamp="2025-10-04 07:22:39 +0000 UTC" firstStartedPulling="2025-10-04 07:22:40.44130279 +0000 UTC m=+777.496201632" lastFinishedPulling="2025-10-04 07:23:25.007510318 +0000 UTC m=+822.062409160" observedRunningTime="2025-10-04 07:23:30.294966171 +0000 UTC m=+827.349865033" watchObservedRunningTime="2025-10-04 07:23:30.686708146 +0000 UTC m=+827.741606988" Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.689277 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2mw6d"] Oct 04 07:23:30 crc kubenswrapper[4987]: I1004 07:23:30.689552 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2mw6d" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="registry-server" containerID="cri-o://ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0" gracePeriod=2 Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.099802 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.248569 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-utilities\") pod \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.248708 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-catalog-content\") pod \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.248759 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d57nh\" (UniqueName: \"kubernetes.io/projected/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-kube-api-access-d57nh\") pod \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\" (UID: \"5f3bd551-83fb-4b4c-98d3-a96dc30119ee\") " Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.249448 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-utilities" (OuterVolumeSpecName: "utilities") pod "5f3bd551-83fb-4b4c-98d3-a96dc30119ee" (UID: "5f3bd551-83fb-4b4c-98d3-a96dc30119ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.254199 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-kube-api-access-d57nh" (OuterVolumeSpecName: "kube-api-access-d57nh") pod "5f3bd551-83fb-4b4c-98d3-a96dc30119ee" (UID: "5f3bd551-83fb-4b4c-98d3-a96dc30119ee"). InnerVolumeSpecName "kube-api-access-d57nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.276729 4987 generic.go:334] "Generic (PLEG): container finished" podID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerID="ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0" exitCode=0 Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.276794 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2mw6d" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.276787 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mw6d" event={"ID":"5f3bd551-83fb-4b4c-98d3-a96dc30119ee","Type":"ContainerDied","Data":"ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0"} Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.278037 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2mw6d" event={"ID":"5f3bd551-83fb-4b4c-98d3-a96dc30119ee","Type":"ContainerDied","Data":"dab9e86a16b835bf41ca09717538f945dbbf70508790749742352850f8444cf4"} Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.278065 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.278080 4987 scope.go:117] "RemoveContainer" containerID="ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.297406 4987 scope.go:117] "RemoveContainer" containerID="bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.301461 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f3bd551-83fb-4b4c-98d3-a96dc30119ee" (UID: "5f3bd551-83fb-4b4c-98d3-a96dc30119ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.313933 4987 scope.go:117] "RemoveContainer" containerID="cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.329331 4987 scope.go:117] "RemoveContainer" containerID="ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0" Oct 04 07:23:31 crc kubenswrapper[4987]: E1004 07:23:31.330021 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0\": container with ID starting with ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0 not found: ID does not exist" containerID="ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.330085 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0"} err="failed to get container status \"ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0\": rpc error: code = NotFound desc = could not find container \"ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0\": container with ID starting with ee7a45fea7337e36c33cc9ecfe69befc44c2dc80db2688d4ab2028d8ad797cc0 not found: ID does not exist" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.330114 4987 scope.go:117] "RemoveContainer" containerID="bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e" Oct 04 07:23:31 crc kubenswrapper[4987]: E1004 07:23:31.330688 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e\": container with ID starting with bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e not found: ID does not exist" containerID="bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.330723 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e"} err="failed to get container status \"bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e\": rpc error: code = NotFound desc = could not find container \"bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e\": container with ID starting with bcab147128a900e4e789400e70cfd37746b7dfcab4d1f8bab723f45008be4d9e not found: ID does not exist" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.330763 4987 scope.go:117] "RemoveContainer" containerID="cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629" Oct 04 07:23:31 crc kubenswrapper[4987]: E1004 07:23:31.331156 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629\": container with ID starting with cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629 not found: ID does not exist" containerID="cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.331200 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629"} err="failed to get container status \"cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629\": rpc error: code = NotFound desc = could not find container \"cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629\": container with ID starting with cc1e4ba137c598c536324627db835fd1da0b972a29a30f47666a3e489ed77629 not found: ID does not exist" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.349929 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.349983 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d57nh\" (UniqueName: \"kubernetes.io/projected/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-kube-api-access-d57nh\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.349995 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3bd551-83fb-4b4c-98d3-a96dc30119ee-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.604571 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2mw6d"] Oct 04 07:23:31 crc kubenswrapper[4987]: I1004 07:23:31.609463 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2mw6d"] Oct 04 07:23:32 crc kubenswrapper[4987]: I1004 07:23:32.062846 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" path="/var/lib/kubelet/pods/5f3bd551-83fb-4b4c-98d3-a96dc30119ee/volumes" Oct 04 07:23:33 crc kubenswrapper[4987]: I1004 07:23:33.455588 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:33 crc kubenswrapper[4987]: I1004 07:23:33.456320 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:33 crc kubenswrapper[4987]: I1004 07:23:33.502250 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:34 crc kubenswrapper[4987]: I1004 07:23:34.342373 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:35 crc kubenswrapper[4987]: I1004 07:23:35.256714 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:23:35 crc kubenswrapper[4987]: I1004 07:23:35.302477 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:23:38 crc kubenswrapper[4987]: I1004 07:23:38.041248 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-n2x7n" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.222692 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-6lqfh" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.265887 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-fdd4g" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.296537 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-67696"] Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.296816 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-67696" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="registry-server" containerID="cri-o://c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e" gracePeriod=2 Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.675386 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.686005 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-catalog-content\") pod \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.686108 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-utilities\") pod \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.686229 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk5jq\" (UniqueName: \"kubernetes.io/projected/8e0fb420-6ff0-412b-9a82-149ea4d941fb-kube-api-access-rk5jq\") pod \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\" (UID: \"8e0fb420-6ff0-412b-9a82-149ea4d941fb\") " Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.687133 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-utilities" (OuterVolumeSpecName: "utilities") pod "8e0fb420-6ff0-412b-9a82-149ea4d941fb" (UID: "8e0fb420-6ff0-412b-9a82-149ea4d941fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.694859 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0fb420-6ff0-412b-9a82-149ea4d941fb-kube-api-access-rk5jq" (OuterVolumeSpecName: "kube-api-access-rk5jq") pod "8e0fb420-6ff0-412b-9a82-149ea4d941fb" (UID: "8e0fb420-6ff0-412b-9a82-149ea4d941fb"). InnerVolumeSpecName "kube-api-access-rk5jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.746211 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e0fb420-6ff0-412b-9a82-149ea4d941fb" (UID: "8e0fb420-6ff0-412b-9a82-149ea4d941fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.787597 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.787655 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk5jq\" (UniqueName: \"kubernetes.io/projected/8e0fb420-6ff0-412b-9a82-149ea4d941fb-kube-api-access-rk5jq\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:40 crc kubenswrapper[4987]: I1004 07:23:40.787673 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e0fb420-6ff0-412b-9a82-149ea4d941fb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.339742 4987 generic.go:334] "Generic (PLEG): container finished" podID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerID="c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e" exitCode=0 Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.339825 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-67696" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.339846 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67696" event={"ID":"8e0fb420-6ff0-412b-9a82-149ea4d941fb","Type":"ContainerDied","Data":"c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e"} Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.340264 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-67696" event={"ID":"8e0fb420-6ff0-412b-9a82-149ea4d941fb","Type":"ContainerDied","Data":"c7259ce2e9c4125442cf50e04b17e35e2b1f9deebd48ad3716bb51498863efc9"} Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.340339 4987 scope.go:117] "RemoveContainer" containerID="c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.357441 4987 scope.go:117] "RemoveContainer" containerID="7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.367899 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-67696"] Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.371304 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-67696"] Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.394086 4987 scope.go:117] "RemoveContainer" containerID="d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.405573 4987 scope.go:117] "RemoveContainer" containerID="c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e" Oct 04 07:23:41 crc kubenswrapper[4987]: E1004 07:23:41.405988 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e\": container with ID starting with c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e not found: ID does not exist" containerID="c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.406026 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e"} err="failed to get container status \"c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e\": rpc error: code = NotFound desc = could not find container \"c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e\": container with ID starting with c5e8cfb54a26f6647dc1e8300887725f7a7cf15ac7ca6404ef01513d549c699e not found: ID does not exist" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.406055 4987 scope.go:117] "RemoveContainer" containerID="7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3" Oct 04 07:23:41 crc kubenswrapper[4987]: E1004 07:23:41.406730 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3\": container with ID starting with 7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3 not found: ID does not exist" containerID="7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.406780 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3"} err="failed to get container status \"7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3\": rpc error: code = NotFound desc = could not find container \"7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3\": container with ID starting with 7cd6d0c8ce82dfa64c5206dfe5471f22f0fa6c9921ca4debc82762778ebac6a3 not found: ID does not exist" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.406814 4987 scope.go:117] "RemoveContainer" containerID="d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264" Oct 04 07:23:41 crc kubenswrapper[4987]: E1004 07:23:41.407081 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264\": container with ID starting with d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264 not found: ID does not exist" containerID="d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264" Oct 04 07:23:41 crc kubenswrapper[4987]: I1004 07:23:41.407105 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264"} err="failed to get container status \"d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264\": rpc error: code = NotFound desc = could not find container \"d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264\": container with ID starting with d017892953a0e536ea3914fccd874a60be82fb0c0ea1195785d5aabc8cb73264 not found: ID does not exist" Oct 04 07:23:42 crc kubenswrapper[4987]: I1004 07:23:42.060920 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" path="/var/lib/kubelet/pods/8e0fb420-6ff0-412b-9a82-149ea4d941fb/volumes" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.267261 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9"] Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269096 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="extract-content" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269165 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="extract-content" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269187 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="extract-utilities" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269199 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="extract-utilities" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269215 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269227 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269242 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="extract-content" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269253 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="extract-content" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269268 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269279 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269296 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="extract-utilities" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269309 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="extract-utilities" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269324 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269335 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269350 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="extract-content" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269361 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="extract-content" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269384 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269395 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: E1004 07:23:45.269408 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="extract-utilities" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269419 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="extract-utilities" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269595 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ecfc35-b0cc-4f19-8fc8-23e659b853e0" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269616 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0fb420-6ff0-412b-9a82-149ea4d941fb" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269660 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="a801564a-9e19-465e-a51a-55c0837f581b" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.269680 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f3bd551-83fb-4b4c-98d3-a96dc30119ee" containerName="registry-server" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.271090 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.274159 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwsts" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.277105 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9"] Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.440820 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-bundle\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.441459 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-util\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.441670 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8v6\" (UniqueName: \"kubernetes.io/projected/31d9f706-030b-4264-bf7f-7847c676a095-kube-api-access-tc8v6\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.543011 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-bundle\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.543128 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-util\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.543173 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8v6\" (UniqueName: \"kubernetes.io/projected/31d9f706-030b-4264-bf7f-7847c676a095-kube-api-access-tc8v6\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.543788 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-bundle\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.544020 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-util\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.568278 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8v6\" (UniqueName: \"kubernetes.io/projected/31d9f706-030b-4264-bf7f-7847c676a095-kube-api-access-tc8v6\") pod \"5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.588586 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:45 crc kubenswrapper[4987]: I1004 07:23:45.982978 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9"] Oct 04 07:23:46 crc kubenswrapper[4987]: I1004 07:23:46.369902 4987 generic.go:334] "Generic (PLEG): container finished" podID="31d9f706-030b-4264-bf7f-7847c676a095" containerID="be4f1447cee0ec8905a3021b54bbf391ad91f5b52abdb9fbc96773fe72617a3e" exitCode=0 Oct 04 07:23:46 crc kubenswrapper[4987]: I1004 07:23:46.370018 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" event={"ID":"31d9f706-030b-4264-bf7f-7847c676a095","Type":"ContainerDied","Data":"be4f1447cee0ec8905a3021b54bbf391ad91f5b52abdb9fbc96773fe72617a3e"} Oct 04 07:23:46 crc kubenswrapper[4987]: I1004 07:23:46.370399 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" event={"ID":"31d9f706-030b-4264-bf7f-7847c676a095","Type":"ContainerStarted","Data":"e2999737d9f10942faf7f5a47cbd6a5cd1f2616efc299d7b7031b84c0d30318a"} Oct 04 07:23:47 crc kubenswrapper[4987]: I1004 07:23:47.382580 4987 generic.go:334] "Generic (PLEG): container finished" podID="31d9f706-030b-4264-bf7f-7847c676a095" containerID="b3f18fda4531c9cb1e3b6f2bd038e1357c26e07956a2e22140b682ad4124d2e2" exitCode=0 Oct 04 07:23:47 crc kubenswrapper[4987]: I1004 07:23:47.382644 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" event={"ID":"31d9f706-030b-4264-bf7f-7847c676a095","Type":"ContainerDied","Data":"b3f18fda4531c9cb1e3b6f2bd038e1357c26e07956a2e22140b682ad4124d2e2"} Oct 04 07:23:48 crc kubenswrapper[4987]: I1004 07:23:48.391210 4987 generic.go:334] "Generic (PLEG): container finished" podID="31d9f706-030b-4264-bf7f-7847c676a095" containerID="5c3c97d832bee423a01e30ee1ec764313d92cf1cb7af7831fd061f5363a8bce6" exitCode=0 Oct 04 07:23:48 crc kubenswrapper[4987]: I1004 07:23:48.391267 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" event={"ID":"31d9f706-030b-4264-bf7f-7847c676a095","Type":"ContainerDied","Data":"5c3c97d832bee423a01e30ee1ec764313d92cf1cb7af7831fd061f5363a8bce6"} Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.646258 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.704459 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-util\") pod \"31d9f706-030b-4264-bf7f-7847c676a095\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.704559 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc8v6\" (UniqueName: \"kubernetes.io/projected/31d9f706-030b-4264-bf7f-7847c676a095-kube-api-access-tc8v6\") pod \"31d9f706-030b-4264-bf7f-7847c676a095\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.704589 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-bundle\") pod \"31d9f706-030b-4264-bf7f-7847c676a095\" (UID: \"31d9f706-030b-4264-bf7f-7847c676a095\") " Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.705574 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-bundle" (OuterVolumeSpecName: "bundle") pod "31d9f706-030b-4264-bf7f-7847c676a095" (UID: "31d9f706-030b-4264-bf7f-7847c676a095"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.710710 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d9f706-030b-4264-bf7f-7847c676a095-kube-api-access-tc8v6" (OuterVolumeSpecName: "kube-api-access-tc8v6") pod "31d9f706-030b-4264-bf7f-7847c676a095" (UID: "31d9f706-030b-4264-bf7f-7847c676a095"). InnerVolumeSpecName "kube-api-access-tc8v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.724936 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-util" (OuterVolumeSpecName: "util") pod "31d9f706-030b-4264-bf7f-7847c676a095" (UID: "31d9f706-030b-4264-bf7f-7847c676a095"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.805438 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc8v6\" (UniqueName: \"kubernetes.io/projected/31d9f706-030b-4264-bf7f-7847c676a095-kube-api-access-tc8v6\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.805473 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:49 crc kubenswrapper[4987]: I1004 07:23:49.805485 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/31d9f706-030b-4264-bf7f-7847c676a095-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:23:50 crc kubenswrapper[4987]: I1004 07:23:50.409769 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" event={"ID":"31d9f706-030b-4264-bf7f-7847c676a095","Type":"ContainerDied","Data":"e2999737d9f10942faf7f5a47cbd6a5cd1f2616efc299d7b7031b84c0d30318a"} Oct 04 07:23:50 crc kubenswrapper[4987]: I1004 07:23:50.409817 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2999737d9f10942faf7f5a47cbd6a5cd1f2616efc299d7b7031b84c0d30318a" Oct 04 07:23:50 crc kubenswrapper[4987]: I1004 07:23:50.409831 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.887118 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n"] Oct 04 07:23:53 crc kubenswrapper[4987]: E1004 07:23:53.888039 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="pull" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.888064 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="pull" Oct 04 07:23:53 crc kubenswrapper[4987]: E1004 07:23:53.888082 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="extract" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.888090 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="extract" Oct 04 07:23:53 crc kubenswrapper[4987]: E1004 07:23:53.888108 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="util" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.888118 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="util" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.888248 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d9f706-030b-4264-bf7f-7847c676a095" containerName="extract" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.889246 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.891617 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-vt7jc" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.892130 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.892169 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.904732 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n"] Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.959730 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72lt5\" (UniqueName: \"kubernetes.io/projected/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-kube-api-access-72lt5\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.959829 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-apiservice-cert\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:53 crc kubenswrapper[4987]: I1004 07:23:53.959960 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-webhook-cert\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.060804 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-webhook-cert\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.060867 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72lt5\" (UniqueName: \"kubernetes.io/projected/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-kube-api-access-72lt5\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.060907 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-apiservice-cert\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.066857 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-webhook-cert\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.067896 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-apiservice-cert\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.078233 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72lt5\" (UniqueName: \"kubernetes.io/projected/bb4722ac-25c0-48c9-86aa-35576a0aa6e5-kube-api-access-72lt5\") pod \"mariadb-operator-controller-manager-6b74947f6c-qsx2n\" (UID: \"bb4722ac-25c0-48c9-86aa-35576a0aa6e5\") " pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.207745 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:23:54 crc kubenswrapper[4987]: I1004 07:23:54.426833 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n"] Oct 04 07:23:55 crc kubenswrapper[4987]: I1004 07:23:55.452045 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" event={"ID":"bb4722ac-25c0-48c9-86aa-35576a0aa6e5","Type":"ContainerStarted","Data":"7540e0c875ca2a6a310ebb8695d3b869b1a467d84adc1253652c69ce3e000754"} Oct 04 07:23:58 crc kubenswrapper[4987]: I1004 07:23:58.480578 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" event={"ID":"bb4722ac-25c0-48c9-86aa-35576a0aa6e5","Type":"ContainerStarted","Data":"2333250e66e527c3265961546006bcec0ef0d337e47d512cdbd691dc7ff73c5e"} Oct 04 07:24:01 crc kubenswrapper[4987]: I1004 07:24:01.500060 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" event={"ID":"bb4722ac-25c0-48c9-86aa-35576a0aa6e5","Type":"ContainerStarted","Data":"edf2fcf84b7d1e4e7dae4b7e5a5f6426bb2ac5efbac9c71c672e14d1e27f1ee5"} Oct 04 07:24:01 crc kubenswrapper[4987]: I1004 07:24:01.500608 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:24:01 crc kubenswrapper[4987]: I1004 07:24:01.521702 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" podStartSLOduration=2.595175162 podStartE2EDuration="8.521681131s" podCreationTimestamp="2025-10-04 07:23:53 +0000 UTC" firstStartedPulling="2025-10-04 07:23:54.439130413 +0000 UTC m=+851.494029255" lastFinishedPulling="2025-10-04 07:24:00.365636382 +0000 UTC m=+857.420535224" observedRunningTime="2025-10-04 07:24:01.520435373 +0000 UTC m=+858.575334215" watchObservedRunningTime="2025-10-04 07:24:01.521681131 +0000 UTC m=+858.576579973" Oct 04 07:24:04 crc kubenswrapper[4987]: I1004 07:24:04.212236 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6b74947f6c-qsx2n" Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.496415 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-rjgsp"] Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.497876 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.500985 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-wlw6l" Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.501648 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-rjgsp"] Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.687494 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jk8l\" (UniqueName: \"kubernetes.io/projected/aeadb69c-fcd3-4354-a0e7-1395d30ce537-kube-api-access-2jk8l\") pod \"infra-operator-index-rjgsp\" (UID: \"aeadb69c-fcd3-4354-a0e7-1395d30ce537\") " pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.788232 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jk8l\" (UniqueName: \"kubernetes.io/projected/aeadb69c-fcd3-4354-a0e7-1395d30ce537-kube-api-access-2jk8l\") pod \"infra-operator-index-rjgsp\" (UID: \"aeadb69c-fcd3-4354-a0e7-1395d30ce537\") " pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.808039 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jk8l\" (UniqueName: \"kubernetes.io/projected/aeadb69c-fcd3-4354-a0e7-1395d30ce537-kube-api-access-2jk8l\") pod \"infra-operator-index-rjgsp\" (UID: \"aeadb69c-fcd3-4354-a0e7-1395d30ce537\") " pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:11 crc kubenswrapper[4987]: I1004 07:24:11.828355 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:12 crc kubenswrapper[4987]: I1004 07:24:12.247304 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-rjgsp"] Oct 04 07:24:12 crc kubenswrapper[4987]: I1004 07:24:12.563616 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-rjgsp" event={"ID":"aeadb69c-fcd3-4354-a0e7-1395d30ce537","Type":"ContainerStarted","Data":"01d574effc8f08dd41f6b9813c4908d4fcf34bb715ace4e35826d3bb73085c54"} Oct 04 07:24:13 crc kubenswrapper[4987]: I1004 07:24:13.570173 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-rjgsp" event={"ID":"aeadb69c-fcd3-4354-a0e7-1395d30ce537","Type":"ContainerStarted","Data":"eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012"} Oct 04 07:24:13 crc kubenswrapper[4987]: I1004 07:24:13.588010 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-rjgsp" podStartSLOduration=1.702518624 podStartE2EDuration="2.587986632s" podCreationTimestamp="2025-10-04 07:24:11 +0000 UTC" firstStartedPulling="2025-10-04 07:24:12.258253786 +0000 UTC m=+869.313152628" lastFinishedPulling="2025-10-04 07:24:13.143721794 +0000 UTC m=+870.198620636" observedRunningTime="2025-10-04 07:24:13.585560515 +0000 UTC m=+870.640459367" watchObservedRunningTime="2025-10-04 07:24:13.587986632 +0000 UTC m=+870.642885474" Oct 04 07:24:15 crc kubenswrapper[4987]: I1004 07:24:15.288079 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-rjgsp"] Oct 04 07:24:15 crc kubenswrapper[4987]: I1004 07:24:15.586719 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-rjgsp" podUID="aeadb69c-fcd3-4354-a0e7-1395d30ce537" containerName="registry-server" containerID="cri-o://eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012" gracePeriod=2 Oct 04 07:24:15 crc kubenswrapper[4987]: I1004 07:24:15.899965 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-mdkmg"] Oct 04 07:24:15 crc kubenswrapper[4987]: I1004 07:24:15.902026 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:15 crc kubenswrapper[4987]: I1004 07:24:15.909322 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-mdkmg"] Oct 04 07:24:15 crc kubenswrapper[4987]: I1004 07:24:15.966790 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.047666 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmg2q\" (UniqueName: \"kubernetes.io/projected/e1582f26-057a-431d-8fff-4e8b0f139149-kube-api-access-hmg2q\") pod \"infra-operator-index-mdkmg\" (UID: \"e1582f26-057a-431d-8fff-4e8b0f139149\") " pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.149041 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jk8l\" (UniqueName: \"kubernetes.io/projected/aeadb69c-fcd3-4354-a0e7-1395d30ce537-kube-api-access-2jk8l\") pod \"aeadb69c-fcd3-4354-a0e7-1395d30ce537\" (UID: \"aeadb69c-fcd3-4354-a0e7-1395d30ce537\") " Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.149361 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmg2q\" (UniqueName: \"kubernetes.io/projected/e1582f26-057a-431d-8fff-4e8b0f139149-kube-api-access-hmg2q\") pod \"infra-operator-index-mdkmg\" (UID: \"e1582f26-057a-431d-8fff-4e8b0f139149\") " pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.157194 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeadb69c-fcd3-4354-a0e7-1395d30ce537-kube-api-access-2jk8l" (OuterVolumeSpecName: "kube-api-access-2jk8l") pod "aeadb69c-fcd3-4354-a0e7-1395d30ce537" (UID: "aeadb69c-fcd3-4354-a0e7-1395d30ce537"). InnerVolumeSpecName "kube-api-access-2jk8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.166196 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmg2q\" (UniqueName: \"kubernetes.io/projected/e1582f26-057a-431d-8fff-4e8b0f139149-kube-api-access-hmg2q\") pod \"infra-operator-index-mdkmg\" (UID: \"e1582f26-057a-431d-8fff-4e8b0f139149\") " pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.251904 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jk8l\" (UniqueName: \"kubernetes.io/projected/aeadb69c-fcd3-4354-a0e7-1395d30ce537-kube-api-access-2jk8l\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.266928 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.551991 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-mdkmg"] Oct 04 07:24:16 crc kubenswrapper[4987]: W1004 07:24:16.558363 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1582f26_057a_431d_8fff_4e8b0f139149.slice/crio-9cc527f3b364ff94c594ed402e8da257cc695ab7a4ece41c0aa4199e42c63370 WatchSource:0}: Error finding container 9cc527f3b364ff94c594ed402e8da257cc695ab7a4ece41c0aa4199e42c63370: Status 404 returned error can't find the container with id 9cc527f3b364ff94c594ed402e8da257cc695ab7a4ece41c0aa4199e42c63370 Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.593241 4987 generic.go:334] "Generic (PLEG): container finished" podID="aeadb69c-fcd3-4354-a0e7-1395d30ce537" containerID="eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012" exitCode=0 Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.593300 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-rjgsp" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.593295 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-rjgsp" event={"ID":"aeadb69c-fcd3-4354-a0e7-1395d30ce537","Type":"ContainerDied","Data":"eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012"} Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.593444 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-rjgsp" event={"ID":"aeadb69c-fcd3-4354-a0e7-1395d30ce537","Type":"ContainerDied","Data":"01d574effc8f08dd41f6b9813c4908d4fcf34bb715ace4e35826d3bb73085c54"} Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.593470 4987 scope.go:117] "RemoveContainer" containerID="eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.594512 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-mdkmg" event={"ID":"e1582f26-057a-431d-8fff-4e8b0f139149","Type":"ContainerStarted","Data":"9cc527f3b364ff94c594ed402e8da257cc695ab7a4ece41c0aa4199e42c63370"} Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.610299 4987 scope.go:117] "RemoveContainer" containerID="eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012" Oct 04 07:24:16 crc kubenswrapper[4987]: E1004 07:24:16.667534 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012\": container with ID starting with eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012 not found: ID does not exist" containerID="eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.667587 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012"} err="failed to get container status \"eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012\": rpc error: code = NotFound desc = could not find container \"eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012\": container with ID starting with eea93b58f93fc328642d7fcc39c5c20721bc52cc7c1e17f86255f0dd079b5012 not found: ID does not exist" Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.677458 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-rjgsp"] Oct 04 07:24:16 crc kubenswrapper[4987]: I1004 07:24:16.680488 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-rjgsp"] Oct 04 07:24:18 crc kubenswrapper[4987]: I1004 07:24:18.062236 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeadb69c-fcd3-4354-a0e7-1395d30ce537" path="/var/lib/kubelet/pods/aeadb69c-fcd3-4354-a0e7-1395d30ce537/volumes" Oct 04 07:24:19 crc kubenswrapper[4987]: I1004 07:24:19.613071 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-mdkmg" event={"ID":"e1582f26-057a-431d-8fff-4e8b0f139149","Type":"ContainerStarted","Data":"41a3ad9f95c8c34d5d0c98b9afac063b94291807c391fe18ce867c8ba9dfdcaf"} Oct 04 07:24:19 crc kubenswrapper[4987]: I1004 07:24:19.627735 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-mdkmg" podStartSLOduration=2.852696182 podStartE2EDuration="4.627717257s" podCreationTimestamp="2025-10-04 07:24:15 +0000 UTC" firstStartedPulling="2025-10-04 07:24:16.561965653 +0000 UTC m=+873.616864495" lastFinishedPulling="2025-10-04 07:24:18.336986728 +0000 UTC m=+875.391885570" observedRunningTime="2025-10-04 07:24:19.626916093 +0000 UTC m=+876.681814935" watchObservedRunningTime="2025-10-04 07:24:19.627717257 +0000 UTC m=+876.682616099" Oct 04 07:24:26 crc kubenswrapper[4987]: I1004 07:24:26.267657 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:26 crc kubenswrapper[4987]: I1004 07:24:26.268949 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:26 crc kubenswrapper[4987]: I1004 07:24:26.293029 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:26 crc kubenswrapper[4987]: I1004 07:24:26.674603 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-mdkmg" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.334772 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp"] Oct 04 07:24:29 crc kubenswrapper[4987]: E1004 07:24:29.335317 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeadb69c-fcd3-4354-a0e7-1395d30ce537" containerName="registry-server" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.335332 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeadb69c-fcd3-4354-a0e7-1395d30ce537" containerName="registry-server" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.335456 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeadb69c-fcd3-4354-a0e7-1395d30ce537" containerName="registry-server" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.336396 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.338661 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwsts" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.344301 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp"] Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.436459 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-bundle\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.436514 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbzpc\" (UniqueName: \"kubernetes.io/projected/d19926e0-f3c3-40e3-96e1-9fac953bdea3-kube-api-access-rbzpc\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.436562 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-util\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.537843 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-util\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.537958 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-bundle\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.537993 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbzpc\" (UniqueName: \"kubernetes.io/projected/d19926e0-f3c3-40e3-96e1-9fac953bdea3-kube-api-access-rbzpc\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.538551 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-util\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.538551 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-bundle\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.559563 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbzpc\" (UniqueName: \"kubernetes.io/projected/d19926e0-f3c3-40e3-96e1-9fac953bdea3-kube-api-access-rbzpc\") pod \"ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:29 crc kubenswrapper[4987]: I1004 07:24:29.678198 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:30 crc kubenswrapper[4987]: I1004 07:24:30.090085 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp"] Oct 04 07:24:30 crc kubenswrapper[4987]: W1004 07:24:30.096604 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd19926e0_f3c3_40e3_96e1_9fac953bdea3.slice/crio-f7da549b80652cce043a2ba8219f2188bdda4e59db60839e5381a4c72b954822 WatchSource:0}: Error finding container f7da549b80652cce043a2ba8219f2188bdda4e59db60839e5381a4c72b954822: Status 404 returned error can't find the container with id f7da549b80652cce043a2ba8219f2188bdda4e59db60839e5381a4c72b954822 Oct 04 07:24:30 crc kubenswrapper[4987]: I1004 07:24:30.671656 4987 generic.go:334] "Generic (PLEG): container finished" podID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerID="18b0a4af39480c337963a042dc902c7ac45d35a9c23880cd3afb9c85835345db" exitCode=0 Oct 04 07:24:30 crc kubenswrapper[4987]: I1004 07:24:30.671698 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" event={"ID":"d19926e0-f3c3-40e3-96e1-9fac953bdea3","Type":"ContainerDied","Data":"18b0a4af39480c337963a042dc902c7ac45d35a9c23880cd3afb9c85835345db"} Oct 04 07:24:30 crc kubenswrapper[4987]: I1004 07:24:30.671723 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" event={"ID":"d19926e0-f3c3-40e3-96e1-9fac953bdea3","Type":"ContainerStarted","Data":"f7da549b80652cce043a2ba8219f2188bdda4e59db60839e5381a4c72b954822"} Oct 04 07:24:32 crc kubenswrapper[4987]: I1004 07:24:32.691774 4987 generic.go:334] "Generic (PLEG): container finished" podID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerID="6c76d01ca35919055e89e2376794676479b5c5c5c3fb22c35a960d638d300ecd" exitCode=0 Oct 04 07:24:32 crc kubenswrapper[4987]: I1004 07:24:32.691916 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" event={"ID":"d19926e0-f3c3-40e3-96e1-9fac953bdea3","Type":"ContainerDied","Data":"6c76d01ca35919055e89e2376794676479b5c5c5c3fb22c35a960d638d300ecd"} Oct 04 07:24:33 crc kubenswrapper[4987]: I1004 07:24:33.699478 4987 generic.go:334] "Generic (PLEG): container finished" podID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerID="35b55448c90932c99461ad892dfb2c4acbb398f4153b3f47cba60679d3c8b13b" exitCode=0 Oct 04 07:24:33 crc kubenswrapper[4987]: I1004 07:24:33.699817 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" event={"ID":"d19926e0-f3c3-40e3-96e1-9fac953bdea3","Type":"ContainerDied","Data":"35b55448c90932c99461ad892dfb2c4acbb398f4153b3f47cba60679d3c8b13b"} Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.107054 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rtlj6"] Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.108297 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.114128 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtlj6"] Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.299754 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn88f\" (UniqueName: \"kubernetes.io/projected/dfe060a7-89fe-4250-9619-76c4f014653e-kube-api-access-nn88f\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.299852 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-catalog-content\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.299933 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-utilities\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.401298 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn88f\" (UniqueName: \"kubernetes.io/projected/dfe060a7-89fe-4250-9619-76c4f014653e-kube-api-access-nn88f\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.401380 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-catalog-content\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.401428 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-utilities\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.401984 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-utilities\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.402060 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-catalog-content\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.424751 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn88f\" (UniqueName: \"kubernetes.io/projected/dfe060a7-89fe-4250-9619-76c4f014653e-kube-api-access-nn88f\") pod \"redhat-marketplace-rtlj6\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.430501 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.818510 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtlj6"] Oct 04 07:24:34 crc kubenswrapper[4987]: I1004 07:24:34.908283 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.010372 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbzpc\" (UniqueName: \"kubernetes.io/projected/d19926e0-f3c3-40e3-96e1-9fac953bdea3-kube-api-access-rbzpc\") pod \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.010439 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-bundle\") pod \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.010474 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-util\") pod \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\" (UID: \"d19926e0-f3c3-40e3-96e1-9fac953bdea3\") " Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.011750 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-bundle" (OuterVolumeSpecName: "bundle") pod "d19926e0-f3c3-40e3-96e1-9fac953bdea3" (UID: "d19926e0-f3c3-40e3-96e1-9fac953bdea3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.017547 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d19926e0-f3c3-40e3-96e1-9fac953bdea3-kube-api-access-rbzpc" (OuterVolumeSpecName: "kube-api-access-rbzpc") pod "d19926e0-f3c3-40e3-96e1-9fac953bdea3" (UID: "d19926e0-f3c3-40e3-96e1-9fac953bdea3"). InnerVolumeSpecName "kube-api-access-rbzpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.026345 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-util" (OuterVolumeSpecName: "util") pod "d19926e0-f3c3-40e3-96e1-9fac953bdea3" (UID: "d19926e0-f3c3-40e3-96e1-9fac953bdea3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.112055 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbzpc\" (UniqueName: \"kubernetes.io/projected/d19926e0-f3c3-40e3-96e1-9fac953bdea3-kube-api-access-rbzpc\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.112092 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.112103 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d19926e0-f3c3-40e3-96e1-9fac953bdea3-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.713665 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" event={"ID":"d19926e0-f3c3-40e3-96e1-9fac953bdea3","Type":"ContainerDied","Data":"f7da549b80652cce043a2ba8219f2188bdda4e59db60839e5381a4c72b954822"} Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.714195 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7da549b80652cce043a2ba8219f2188bdda4e59db60839e5381a4c72b954822" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.714293 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp" Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.718788 4987 generic.go:334] "Generic (PLEG): container finished" podID="dfe060a7-89fe-4250-9619-76c4f014653e" containerID="f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073" exitCode=0 Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.718863 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtlj6" event={"ID":"dfe060a7-89fe-4250-9619-76c4f014653e","Type":"ContainerDied","Data":"f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073"} Oct 04 07:24:35 crc kubenswrapper[4987]: I1004 07:24:35.718917 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtlj6" event={"ID":"dfe060a7-89fe-4250-9619-76c4f014653e","Type":"ContainerStarted","Data":"2abfa1234afe26a4d359e74e5f0e978d6468651a75c9cc274f571f9eb721a271"} Oct 04 07:24:36 crc kubenswrapper[4987]: I1004 07:24:36.734878 4987 generic.go:334] "Generic (PLEG): container finished" podID="dfe060a7-89fe-4250-9619-76c4f014653e" containerID="8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323" exitCode=0 Oct 04 07:24:36 crc kubenswrapper[4987]: I1004 07:24:36.734973 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtlj6" event={"ID":"dfe060a7-89fe-4250-9619-76c4f014653e","Type":"ContainerDied","Data":"8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323"} Oct 04 07:24:37 crc kubenswrapper[4987]: I1004 07:24:37.757281 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtlj6" event={"ID":"dfe060a7-89fe-4250-9619-76c4f014653e","Type":"ContainerStarted","Data":"5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2"} Oct 04 07:24:37 crc kubenswrapper[4987]: I1004 07:24:37.778874 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rtlj6" podStartSLOduration=2.357826477 podStartE2EDuration="3.77883964s" podCreationTimestamp="2025-10-04 07:24:34 +0000 UTC" firstStartedPulling="2025-10-04 07:24:35.720894007 +0000 UTC m=+892.775792849" lastFinishedPulling="2025-10-04 07:24:37.14190717 +0000 UTC m=+894.196806012" observedRunningTime="2025-10-04 07:24:37.778151069 +0000 UTC m=+894.833049921" watchObservedRunningTime="2025-10-04 07:24:37.77883964 +0000 UTC m=+894.833738482" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.159436 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv"] Oct 04 07:24:43 crc kubenswrapper[4987]: E1004 07:24:43.160175 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="util" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.160189 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="util" Oct 04 07:24:43 crc kubenswrapper[4987]: E1004 07:24:43.160209 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="extract" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.160215 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="extract" Oct 04 07:24:43 crc kubenswrapper[4987]: E1004 07:24:43.160226 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="pull" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.160232 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="pull" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.160341 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="d19926e0-f3c3-40e3-96e1-9fac953bdea3" containerName="extract" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.161005 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.163604 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.164673 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6dz65" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.178864 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv"] Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.344441 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c819f4f-118f-4592-8f51-b2f564437526-apiservice-cert\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.344533 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zgf4\" (UniqueName: \"kubernetes.io/projected/1c819f4f-118f-4592-8f51-b2f564437526-kube-api-access-4zgf4\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.344553 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c819f4f-118f-4592-8f51-b2f564437526-webhook-cert\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.445976 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c819f4f-118f-4592-8f51-b2f564437526-webhook-cert\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.446093 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c819f4f-118f-4592-8f51-b2f564437526-apiservice-cert\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.446215 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zgf4\" (UniqueName: \"kubernetes.io/projected/1c819f4f-118f-4592-8f51-b2f564437526-kube-api-access-4zgf4\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.453361 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c819f4f-118f-4592-8f51-b2f564437526-apiservice-cert\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.453381 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c819f4f-118f-4592-8f51-b2f564437526-webhook-cert\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.466250 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zgf4\" (UniqueName: \"kubernetes.io/projected/1c819f4f-118f-4592-8f51-b2f564437526-kube-api-access-4zgf4\") pod \"infra-operator-controller-manager-759bcdfbb8-g4mlv\" (UID: \"1c819f4f-118f-4592-8f51-b2f564437526\") " pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.490091 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.753614 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv"] Oct 04 07:24:43 crc kubenswrapper[4987]: W1004 07:24:43.754542 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c819f4f_118f_4592_8f51_b2f564437526.slice/crio-667664e5aff6b291d7dae468df33d3a3e66c894e43aa3b6b50243859dcd6e61a WatchSource:0}: Error finding container 667664e5aff6b291d7dae468df33d3a3e66c894e43aa3b6b50243859dcd6e61a: Status 404 returned error can't find the container with id 667664e5aff6b291d7dae468df33d3a3e66c894e43aa3b6b50243859dcd6e61a Oct 04 07:24:43 crc kubenswrapper[4987]: I1004 07:24:43.795923 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" event={"ID":"1c819f4f-118f-4592-8f51-b2f564437526","Type":"ContainerStarted","Data":"667664e5aff6b291d7dae468df33d3a3e66c894e43aa3b6b50243859dcd6e61a"} Oct 04 07:24:44 crc kubenswrapper[4987]: I1004 07:24:44.431368 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:44 crc kubenswrapper[4987]: I1004 07:24:44.431802 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:44 crc kubenswrapper[4987]: I1004 07:24:44.476891 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:44 crc kubenswrapper[4987]: I1004 07:24:44.858478 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:45 crc kubenswrapper[4987]: I1004 07:24:45.808760 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" event={"ID":"1c819f4f-118f-4592-8f51-b2f564437526","Type":"ContainerStarted","Data":"507912c9afeb3b8f06ec52ec006362349b86fa9c975d1ee0979abb02436bcd09"} Oct 04 07:24:45 crc kubenswrapper[4987]: I1004 07:24:45.809133 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" event={"ID":"1c819f4f-118f-4592-8f51-b2f564437526","Type":"ContainerStarted","Data":"ded568f9f20374c86304247b794ecb6c0f3a04412808268225b262b4ba544c7d"} Oct 04 07:24:45 crc kubenswrapper[4987]: I1004 07:24:45.809207 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:45 crc kubenswrapper[4987]: I1004 07:24:45.830087 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" podStartSLOduration=1.29260579 podStartE2EDuration="2.830061561s" podCreationTimestamp="2025-10-04 07:24:43 +0000 UTC" firstStartedPulling="2025-10-04 07:24:43.756872149 +0000 UTC m=+900.811770991" lastFinishedPulling="2025-10-04 07:24:45.29432792 +0000 UTC m=+902.349226762" observedRunningTime="2025-10-04 07:24:45.827540201 +0000 UTC m=+902.882439053" watchObservedRunningTime="2025-10-04 07:24:45.830061561 +0000 UTC m=+902.884960423" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.089207 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtlj6"] Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.089770 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rtlj6" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="registry-server" containerID="cri-o://5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2" gracePeriod=2 Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.444197 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.620021 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-utilities\") pod \"dfe060a7-89fe-4250-9619-76c4f014653e\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.620378 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-catalog-content\") pod \"dfe060a7-89fe-4250-9619-76c4f014653e\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.620482 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn88f\" (UniqueName: \"kubernetes.io/projected/dfe060a7-89fe-4250-9619-76c4f014653e-kube-api-access-nn88f\") pod \"dfe060a7-89fe-4250-9619-76c4f014653e\" (UID: \"dfe060a7-89fe-4250-9619-76c4f014653e\") " Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.621163 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-utilities" (OuterVolumeSpecName: "utilities") pod "dfe060a7-89fe-4250-9619-76c4f014653e" (UID: "dfe060a7-89fe-4250-9619-76c4f014653e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.626970 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe060a7-89fe-4250-9619-76c4f014653e-kube-api-access-nn88f" (OuterVolumeSpecName: "kube-api-access-nn88f") pod "dfe060a7-89fe-4250-9619-76c4f014653e" (UID: "dfe060a7-89fe-4250-9619-76c4f014653e"). InnerVolumeSpecName "kube-api-access-nn88f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.636331 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfe060a7-89fe-4250-9619-76c4f014653e" (UID: "dfe060a7-89fe-4250-9619-76c4f014653e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.721610 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn88f\" (UniqueName: \"kubernetes.io/projected/dfe060a7-89fe-4250-9619-76c4f014653e-kube-api-access-nn88f\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.721882 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.721965 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe060a7-89fe-4250-9619-76c4f014653e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.833096 4987 generic.go:334] "Generic (PLEG): container finished" podID="dfe060a7-89fe-4250-9619-76c4f014653e" containerID="5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2" exitCode=0 Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.833148 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtlj6" event={"ID":"dfe060a7-89fe-4250-9619-76c4f014653e","Type":"ContainerDied","Data":"5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2"} Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.833186 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rtlj6" event={"ID":"dfe060a7-89fe-4250-9619-76c4f014653e","Type":"ContainerDied","Data":"2abfa1234afe26a4d359e74e5f0e978d6468651a75c9cc274f571f9eb721a271"} Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.833193 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rtlj6" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.833206 4987 scope.go:117] "RemoveContainer" containerID="5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.861926 4987 scope.go:117] "RemoveContainer" containerID="8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.872527 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtlj6"] Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.882824 4987 scope.go:117] "RemoveContainer" containerID="f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.890206 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rtlj6"] Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.905600 4987 scope.go:117] "RemoveContainer" containerID="5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2" Oct 04 07:24:48 crc kubenswrapper[4987]: E1004 07:24:48.906372 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2\": container with ID starting with 5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2 not found: ID does not exist" containerID="5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.906411 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2"} err="failed to get container status \"5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2\": rpc error: code = NotFound desc = could not find container \"5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2\": container with ID starting with 5846bb421fffab65e995ff0d583ed2a4750d6d309fa9917fcf1e60a2c65eeca2 not found: ID does not exist" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.906437 4987 scope.go:117] "RemoveContainer" containerID="8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323" Oct 04 07:24:48 crc kubenswrapper[4987]: E1004 07:24:48.906817 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323\": container with ID starting with 8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323 not found: ID does not exist" containerID="8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.906844 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323"} err="failed to get container status \"8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323\": rpc error: code = NotFound desc = could not find container \"8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323\": container with ID starting with 8e7014d684ca295207f1746b6a2f95a3cb061ca4cbe09cc79d2defed1298a323 not found: ID does not exist" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.906861 4987 scope.go:117] "RemoveContainer" containerID="f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073" Oct 04 07:24:48 crc kubenswrapper[4987]: E1004 07:24:48.907166 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073\": container with ID starting with f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073 not found: ID does not exist" containerID="f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073" Oct 04 07:24:48 crc kubenswrapper[4987]: I1004 07:24:48.907208 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073"} err="failed to get container status \"f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073\": rpc error: code = NotFound desc = could not find container \"f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073\": container with ID starting with f90965c21bf45791516a6287495db34e15788e1c5ab3af17f8833b8bda04e073 not found: ID does not exist" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.827548 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 04 07:24:49 crc kubenswrapper[4987]: E1004 07:24:49.827820 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="extract-content" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.827834 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="extract-content" Oct 04 07:24:49 crc kubenswrapper[4987]: E1004 07:24:49.827847 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="registry-server" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.827853 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="registry-server" Oct 04 07:24:49 crc kubenswrapper[4987]: E1004 07:24:49.827915 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="extract-utilities" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.827923 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="extract-utilities" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.828025 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" containerName="registry-server" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.828644 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.830539 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.830859 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.830999 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.831001 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.830999 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-lhtk4" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.831393 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.840683 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.848421 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.849617 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.855223 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.856215 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.872657 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.891574 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.937230 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d7b90541-0826-4c5c-a22a-ebebb127e444-secrets\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.937508 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6tpl\" (UniqueName: \"kubernetes.io/projected/d7b90541-0826-4c5c-a22a-ebebb127e444-kube-api-access-q6tpl\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.937643 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d7b90541-0826-4c5c-a22a-ebebb127e444-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.937734 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-config-data-default\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.937855 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.937960 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:49 crc kubenswrapper[4987]: I1004 07:24:49.938073 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-kolla-config\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039045 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d6684ada-c0e3-4f8f-a804-80f711240730-secrets\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039099 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62mt\" (UniqueName: \"kubernetes.io/projected/d5ddfd58-951f-46d4-804b-0ae9998befc1-kube-api-access-d62mt\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039138 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d7b90541-0826-4c5c-a22a-ebebb127e444-secrets\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039164 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039198 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6tpl\" (UniqueName: \"kubernetes.io/projected/d7b90541-0826-4c5c-a22a-ebebb127e444-kube-api-access-q6tpl\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039331 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-config-data-default\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039422 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d7b90541-0826-4c5c-a22a-ebebb127e444-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039460 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-config-data-default\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039507 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039539 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039566 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-config-data-default\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039602 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-operator-scripts\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039678 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-kolla-config\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039702 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-operator-scripts\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039753 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-kolla-config\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039777 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6684ada-c0e3-4f8f-a804-80f711240730-config-data-generated\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039818 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d5ddfd58-951f-46d4-804b-0ae9998befc1-config-data-generated\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039887 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2vjv\" (UniqueName: \"kubernetes.io/projected/d6684ada-c0e3-4f8f-a804-80f711240730-kube-api-access-f2vjv\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039915 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-kolla-config\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039938 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d5ddfd58-951f-46d4-804b-0ae9998befc1-secrets\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.039962 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.040531 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.040834 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d7b90541-0826-4c5c-a22a-ebebb127e444-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.041563 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-config-data-default\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.042260 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.042655 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d7b90541-0826-4c5c-a22a-ebebb127e444-kolla-config\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.044369 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d7b90541-0826-4c5c-a22a-ebebb127e444-secrets\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.056205 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6tpl\" (UniqueName: \"kubernetes.io/projected/d7b90541-0826-4c5c-a22a-ebebb127e444-kube-api-access-q6tpl\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.059096 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"d7b90541-0826-4c5c-a22a-ebebb127e444\") " pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.061485 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe060a7-89fe-4250-9619-76c4f014653e" path="/var/lib/kubelet/pods/dfe060a7-89fe-4250-9619-76c4f014653e/volumes" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.141775 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d6684ada-c0e3-4f8f-a804-80f711240730-secrets\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.141837 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62mt\" (UniqueName: \"kubernetes.io/projected/d5ddfd58-951f-46d4-804b-0ae9998befc1-kube-api-access-d62mt\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.141865 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.141903 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-config-data-default\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.141977 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-config-data-default\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.142003 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-operator-scripts\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.142049 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-kolla-config\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.142075 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-operator-scripts\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.142103 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6684ada-c0e3-4f8f-a804-80f711240730-config-data-generated\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.142131 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.142849 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-config-data-default\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.143416 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.143707 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d5ddfd58-951f-46d4-804b-0ae9998befc1-config-data-generated\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.143845 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-kolla-config\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.143867 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2vjv\" (UniqueName: \"kubernetes.io/projected/d6684ada-c0e3-4f8f-a804-80f711240730-kube-api-access-f2vjv\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.143894 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d5ddfd58-951f-46d4-804b-0ae9998befc1-secrets\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.143919 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.144032 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d5ddfd58-951f-46d4-804b-0ae9998befc1-config-data-generated\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.144202 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.144392 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-kolla-config\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.144944 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d6684ada-c0e3-4f8f-a804-80f711240730-config-data-generated\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.145542 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d6684ada-c0e3-4f8f-a804-80f711240730-secrets\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.146828 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-kolla-config\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.147087 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-operator-scripts\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.147460 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5ddfd58-951f-46d4-804b-0ae9998befc1-operator-scripts\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.147697 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d6684ada-c0e3-4f8f-a804-80f711240730-config-data-default\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.148224 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d5ddfd58-951f-46d4-804b-0ae9998befc1-secrets\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.159289 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62mt\" (UniqueName: \"kubernetes.io/projected/d5ddfd58-951f-46d4-804b-0ae9998befc1-kube-api-access-d62mt\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.160993 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.162259 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2vjv\" (UniqueName: \"kubernetes.io/projected/d6684ada-c0e3-4f8f-a804-80f711240730-kube-api-access-f2vjv\") pod \"openstack-galera-1\" (UID: \"d6684ada-c0e3-4f8f-a804-80f711240730\") " pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.166125 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-2\" (UID: \"d5ddfd58-951f-46d4-804b-0ae9998befc1\") " pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.197669 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.469616 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.586138 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.624342 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.847685 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"d6684ada-c0e3-4f8f-a804-80f711240730","Type":"ContainerStarted","Data":"f52cb10db45c37e8911be5230af8d507ff799dbf40d9ed927fce4bade424c4db"} Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.849111 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"d7b90541-0826-4c5c-a22a-ebebb127e444","Type":"ContainerStarted","Data":"bac20f409016a9203c9f6dffd8a0a1a1ff3dfec020480d01da0d24174de1168f"} Oct 04 07:24:50 crc kubenswrapper[4987]: I1004 07:24:50.925673 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Oct 04 07:24:50 crc kubenswrapper[4987]: W1004 07:24:50.930882 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5ddfd58_951f_46d4_804b_0ae9998befc1.slice/crio-44b5889243b7ea344bbac94343b28f40edc4d3a96ab3590a6450a86ffed8019d WatchSource:0}: Error finding container 44b5889243b7ea344bbac94343b28f40edc4d3a96ab3590a6450a86ffed8019d: Status 404 returned error can't find the container with id 44b5889243b7ea344bbac94343b28f40edc4d3a96ab3590a6450a86ffed8019d Oct 04 07:24:51 crc kubenswrapper[4987]: I1004 07:24:51.859905 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d5ddfd58-951f-46d4-804b-0ae9998befc1","Type":"ContainerStarted","Data":"44b5889243b7ea344bbac94343b28f40edc4d3a96ab3590a6450a86ffed8019d"} Oct 04 07:24:53 crc kubenswrapper[4987]: I1004 07:24:53.496102 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-759bcdfbb8-g4mlv" Oct 04 07:24:57 crc kubenswrapper[4987]: I1004 07:24:57.911404 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-j9wbv"] Oct 04 07:24:57 crc kubenswrapper[4987]: I1004 07:24:57.919012 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:24:57 crc kubenswrapper[4987]: I1004 07:24:57.930084 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-29z2h" Oct 04 07:24:57 crc kubenswrapper[4987]: I1004 07:24:57.933357 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-j9wbv"] Oct 04 07:24:58 crc kubenswrapper[4987]: I1004 07:24:58.081668 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx5sh\" (UniqueName: \"kubernetes.io/projected/3dc91b62-1cd4-40db-b853-ff83dd36cd0d-kube-api-access-lx5sh\") pod \"rabbitmq-cluster-operator-index-j9wbv\" (UID: \"3dc91b62-1cd4-40db-b853-ff83dd36cd0d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:24:58 crc kubenswrapper[4987]: I1004 07:24:58.183221 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx5sh\" (UniqueName: \"kubernetes.io/projected/3dc91b62-1cd4-40db-b853-ff83dd36cd0d-kube-api-access-lx5sh\") pod \"rabbitmq-cluster-operator-index-j9wbv\" (UID: \"3dc91b62-1cd4-40db-b853-ff83dd36cd0d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:24:58 crc kubenswrapper[4987]: I1004 07:24:58.207506 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx5sh\" (UniqueName: \"kubernetes.io/projected/3dc91b62-1cd4-40db-b853-ff83dd36cd0d-kube-api-access-lx5sh\") pod \"rabbitmq-cluster-operator-index-j9wbv\" (UID: \"3dc91b62-1cd4-40db-b853-ff83dd36cd0d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:24:58 crc kubenswrapper[4987]: I1004 07:24:58.243891 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.213251 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.214346 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.216593 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-p98p5" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.217017 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.228846 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.316068 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d20ac66d-0466-4478-af96-72fc6adc11d8-kolla-config\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.316417 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d20ac66d-0466-4478-af96-72fc6adc11d8-config-data\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.316613 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-875pd\" (UniqueName: \"kubernetes.io/projected/d20ac66d-0466-4478-af96-72fc6adc11d8-kube-api-access-875pd\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.418105 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d20ac66d-0466-4478-af96-72fc6adc11d8-kolla-config\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.418158 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d20ac66d-0466-4478-af96-72fc6adc11d8-config-data\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.418182 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-875pd\" (UniqueName: \"kubernetes.io/projected/d20ac66d-0466-4478-af96-72fc6adc11d8-kube-api-access-875pd\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.419016 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d20ac66d-0466-4478-af96-72fc6adc11d8-kolla-config\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.419308 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d20ac66d-0466-4478-af96-72fc6adc11d8-config-data\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.438475 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-875pd\" (UniqueName: \"kubernetes.io/projected/d20ac66d-0466-4478-af96-72fc6adc11d8-kube-api-access-875pd\") pod \"memcached-0\" (UID: \"d20ac66d-0466-4478-af96-72fc6adc11d8\") " pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:00 crc kubenswrapper[4987]: I1004 07:25:00.531885 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:03 crc kubenswrapper[4987]: I1004 07:25:03.034553 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Oct 04 07:25:03 crc kubenswrapper[4987]: W1004 07:25:03.041027 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd20ac66d_0466_4478_af96_72fc6adc11d8.slice/crio-43e7962e6ff87fed7fc9d1134a304a6effdc5ac4fcdc996d87bb25ad299fb1b0 WatchSource:0}: Error finding container 43e7962e6ff87fed7fc9d1134a304a6effdc5ac4fcdc996d87bb25ad299fb1b0: Status 404 returned error can't find the container with id 43e7962e6ff87fed7fc9d1134a304a6effdc5ac4fcdc996d87bb25ad299fb1b0 Oct 04 07:25:03 crc kubenswrapper[4987]: I1004 07:25:03.285806 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-j9wbv"] Oct 04 07:25:03 crc kubenswrapper[4987]: E1004 07:25:03.851029 4987 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:5d10c016b13499110b5f9ca2bccfaf6d2fd4298c9f02580d7208fe91850da0a6" Oct 04 07:25:03 crc kubenswrapper[4987]: E1004 07:25:03.851545 4987 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:5d10c016b13499110b5f9ca2bccfaf6d2fd4298c9f02580d7208fe91850da0a6,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:DB_ROOT_PASSWORD,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:DbRootPassword,Optional:nil,},},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secrets,ReadOnly:true,MountPath:/var/lib/secrets,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q6tpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_glance-kuttl-tests(d7b90541-0826-4c5c-a22a-ebebb127e444): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 07:25:03 crc kubenswrapper[4987]: E1004 07:25:03.852796 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/openstack-galera-0" podUID="d7b90541-0826-4c5c-a22a-ebebb127e444" Oct 04 07:25:03 crc kubenswrapper[4987]: I1004 07:25:03.962938 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" event={"ID":"3dc91b62-1cd4-40db-b853-ff83dd36cd0d","Type":"ContainerStarted","Data":"b115a82aeba0a8d2b1dc9002cccec3f049f80202d876573a6ff268515acbcdf8"} Oct 04 07:25:03 crc kubenswrapper[4987]: I1004 07:25:03.964917 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"d20ac66d-0466-4478-af96-72fc6adc11d8","Type":"ContainerStarted","Data":"43e7962e6ff87fed7fc9d1134a304a6effdc5ac4fcdc996d87bb25ad299fb1b0"} Oct 04 07:25:04 crc kubenswrapper[4987]: I1004 07:25:04.974401 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d5ddfd58-951f-46d4-804b-0ae9998befc1","Type":"ContainerStarted","Data":"0023c70f3c6df38af101adeaa4a195cb2e9f2f26b17b8fbc4ba5924e1179f1b3"} Oct 04 07:25:04 crc kubenswrapper[4987]: I1004 07:25:04.977671 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"d6684ada-c0e3-4f8f-a804-80f711240730","Type":"ContainerStarted","Data":"87a6f7c74e01641f272e3fd432b58c3b07fb2e5d52f8abc5928c0fb97d229dce"} Oct 04 07:25:04 crc kubenswrapper[4987]: I1004 07:25:04.980316 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"d7b90541-0826-4c5c-a22a-ebebb127e444","Type":"ContainerStarted","Data":"84b8c0222b2a5000d7a1dab932090979f5d9e03e9a97ca14c422a06d38b3a3e3"} Oct 04 07:25:08 crc kubenswrapper[4987]: I1004 07:25:08.003046 4987 generic.go:334] "Generic (PLEG): container finished" podID="d6684ada-c0e3-4f8f-a804-80f711240730" containerID="87a6f7c74e01641f272e3fd432b58c3b07fb2e5d52f8abc5928c0fb97d229dce" exitCode=0 Oct 04 07:25:08 crc kubenswrapper[4987]: I1004 07:25:08.003259 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"d6684ada-c0e3-4f8f-a804-80f711240730","Type":"ContainerDied","Data":"87a6f7c74e01641f272e3fd432b58c3b07fb2e5d52f8abc5928c0fb97d229dce"} Oct 04 07:25:08 crc kubenswrapper[4987]: I1004 07:25:08.005852 4987 generic.go:334] "Generic (PLEG): container finished" podID="d5ddfd58-951f-46d4-804b-0ae9998befc1" containerID="0023c70f3c6df38af101adeaa4a195cb2e9f2f26b17b8fbc4ba5924e1179f1b3" exitCode=0 Oct 04 07:25:08 crc kubenswrapper[4987]: I1004 07:25:08.005904 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d5ddfd58-951f-46d4-804b-0ae9998befc1","Type":"ContainerDied","Data":"0023c70f3c6df38af101adeaa4a195cb2e9f2f26b17b8fbc4ba5924e1179f1b3"} Oct 04 07:25:09 crc kubenswrapper[4987]: I1004 07:25:09.012790 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"d20ac66d-0466-4478-af96-72fc6adc11d8","Type":"ContainerStarted","Data":"d3e1879253e1d4cca53293545fa552133821ad6bbf2fb6661248aaab175ec76c"} Oct 04 07:25:09 crc kubenswrapper[4987]: I1004 07:25:09.013981 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"d5ddfd58-951f-46d4-804b-0ae9998befc1","Type":"ContainerStarted","Data":"137a5bfd19e6f3f65e7b6291524bd3b96b478289816abc428527ee9580be8854"} Oct 04 07:25:09 crc kubenswrapper[4987]: I1004 07:25:09.019853 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"d6684ada-c0e3-4f8f-a804-80f711240730","Type":"ContainerStarted","Data":"4610055cb4edd0ff90781b71355dbcf077168199d9ef105ff477aa5614750252"} Oct 04 07:25:09 crc kubenswrapper[4987]: I1004 07:25:09.021162 4987 generic.go:334] "Generic (PLEG): container finished" podID="d7b90541-0826-4c5c-a22a-ebebb127e444" containerID="84b8c0222b2a5000d7a1dab932090979f5d9e03e9a97ca14c422a06d38b3a3e3" exitCode=0 Oct 04 07:25:09 crc kubenswrapper[4987]: I1004 07:25:09.021212 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"d7b90541-0826-4c5c-a22a-ebebb127e444","Type":"ContainerDied","Data":"84b8c0222b2a5000d7a1dab932090979f5d9e03e9a97ca14c422a06d38b3a3e3"} Oct 04 07:25:09 crc kubenswrapper[4987]: I1004 07:25:09.090068 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=8.208661749000001 podStartE2EDuration="21.090050076s" podCreationTimestamp="2025-10-04 07:24:48 +0000 UTC" firstStartedPulling="2025-10-04 07:24:50.933693629 +0000 UTC m=+907.988592471" lastFinishedPulling="2025-10-04 07:25:03.815081956 +0000 UTC m=+920.869980798" observedRunningTime="2025-10-04 07:25:09.058815035 +0000 UTC m=+926.113713877" watchObservedRunningTime="2025-10-04 07:25:09.090050076 +0000 UTC m=+926.144948918" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.029471 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" event={"ID":"3dc91b62-1cd4-40db-b853-ff83dd36cd0d","Type":"ContainerStarted","Data":"3b3eddc47e8fa922532a1153d40984ef6a6e97226b8d59b612336c8d961df49d"} Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.033303 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"d7b90541-0826-4c5c-a22a-ebebb127e444","Type":"ContainerStarted","Data":"d870255dd32ae3b425590dadd09bf6f6630ecfca44eea21f2f2107a504a106a6"} Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.034207 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.051851 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" podStartSLOduration=7.683345842 podStartE2EDuration="13.051832812s" podCreationTimestamp="2025-10-04 07:24:57 +0000 UTC" firstStartedPulling="2025-10-04 07:25:03.735957619 +0000 UTC m=+920.790856451" lastFinishedPulling="2025-10-04 07:25:09.104444579 +0000 UTC m=+926.159343421" observedRunningTime="2025-10-04 07:25:10.049087446 +0000 UTC m=+927.103986288" watchObservedRunningTime="2025-10-04 07:25:10.051832812 +0000 UTC m=+927.106731654" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.075283 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=-9223372014.779512 podStartE2EDuration="22.075262648s" podCreationTimestamp="2025-10-04 07:24:48 +0000 UTC" firstStartedPulling="2025-10-04 07:24:50.59570357 +0000 UTC m=+907.650602432" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:25:10.069305391 +0000 UTC m=+927.124204233" watchObservedRunningTime="2025-10-04 07:25:10.075262648 +0000 UTC m=+927.130161490" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.090178 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=6.178877636 podStartE2EDuration="10.090161285s" podCreationTimestamp="2025-10-04 07:25:00 +0000 UTC" firstStartedPulling="2025-10-04 07:25:03.043804085 +0000 UTC m=+920.098702927" lastFinishedPulling="2025-10-04 07:25:06.955087734 +0000 UTC m=+924.009986576" observedRunningTime="2025-10-04 07:25:10.086334946 +0000 UTC m=+927.141233798" watchObservedRunningTime="2025-10-04 07:25:10.090161285 +0000 UTC m=+927.145060127" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.111599 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=8.982247003 podStartE2EDuration="22.111580459s" podCreationTimestamp="2025-10-04 07:24:48 +0000 UTC" firstStartedPulling="2025-10-04 07:24:50.636052678 +0000 UTC m=+907.690951520" lastFinishedPulling="2025-10-04 07:25:03.765386144 +0000 UTC m=+920.820284976" observedRunningTime="2025-10-04 07:25:10.103105663 +0000 UTC m=+927.158004515" watchObservedRunningTime="2025-10-04 07:25:10.111580459 +0000 UTC m=+927.166479301" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.144351 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.144428 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.198747 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.198850 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.470454 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:25:10 crc kubenswrapper[4987]: I1004 07:25:10.470525 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:25:12 crc kubenswrapper[4987]: I1004 07:25:12.703735 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:25:12 crc kubenswrapper[4987]: I1004 07:25:12.704017 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:25:15 crc kubenswrapper[4987]: I1004 07:25:15.533932 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Oct 04 07:25:16 crc kubenswrapper[4987]: I1004 07:25:16.535925 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:25:16 crc kubenswrapper[4987]: I1004 07:25:16.578712 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Oct 04 07:25:18 crc kubenswrapper[4987]: I1004 07:25:18.244653 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:25:18 crc kubenswrapper[4987]: I1004 07:25:18.244712 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:25:18 crc kubenswrapper[4987]: I1004 07:25:18.280337 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:25:19 crc kubenswrapper[4987]: I1004 07:25:19.136243 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-j9wbv" Oct 04 07:25:20 crc kubenswrapper[4987]: I1004 07:25:20.509028 4987 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="d5ddfd58-951f-46d4-804b-0ae9998befc1" containerName="galera" probeResult="failure" output=< Oct 04 07:25:20 crc kubenswrapper[4987]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Oct 04 07:25:20 crc kubenswrapper[4987]: > Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.528152 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2"] Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.529758 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.531793 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwsts" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.548754 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2"] Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.686333 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.686469 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.686524 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tscjx\" (UniqueName: \"kubernetes.io/projected/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-kube-api-access-tscjx\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.788378 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.788524 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.789034 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.789060 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tscjx\" (UniqueName: \"kubernetes.io/projected/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-kube-api-access-tscjx\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.789034 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.806943 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tscjx\" (UniqueName: \"kubernetes.io/projected/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-kube-api-access-tscjx\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:28 crc kubenswrapper[4987]: I1004 07:25:28.847779 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:29 crc kubenswrapper[4987]: I1004 07:25:29.262691 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2"] Oct 04 07:25:29 crc kubenswrapper[4987]: I1004 07:25:29.615252 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:25:29 crc kubenswrapper[4987]: I1004 07:25:29.666799 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Oct 04 07:25:30 crc kubenswrapper[4987]: I1004 07:25:30.168376 4987 generic.go:334] "Generic (PLEG): container finished" podID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerID="e53dc71c905b4b2d7d50989f343c042221e034973f9c541b3f7df71fc9ed7375" exitCode=0 Oct 04 07:25:30 crc kubenswrapper[4987]: I1004 07:25:30.168488 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" event={"ID":"82cd04d6-e9cc-4f3a-89f6-d867a256a4af","Type":"ContainerDied","Data":"e53dc71c905b4b2d7d50989f343c042221e034973f9c541b3f7df71fc9ed7375"} Oct 04 07:25:30 crc kubenswrapper[4987]: I1004 07:25:30.168840 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" event={"ID":"82cd04d6-e9cc-4f3a-89f6-d867a256a4af","Type":"ContainerStarted","Data":"76acf1f995c1ca499b6356d67c5fd48726bdfbcad6d3dd1fcada0cd3b429d9ff"} Oct 04 07:25:30 crc kubenswrapper[4987]: I1004 07:25:30.572831 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:25:30 crc kubenswrapper[4987]: I1004 07:25:30.626897 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Oct 04 07:25:31 crc kubenswrapper[4987]: I1004 07:25:31.178329 4987 generic.go:334] "Generic (PLEG): container finished" podID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerID="7afd436533560c218ede42dd6c56f50e1ca916c19826b3393856f0fc5f6cef1f" exitCode=0 Oct 04 07:25:31 crc kubenswrapper[4987]: I1004 07:25:31.178431 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" event={"ID":"82cd04d6-e9cc-4f3a-89f6-d867a256a4af","Type":"ContainerDied","Data":"7afd436533560c218ede42dd6c56f50e1ca916c19826b3393856f0fc5f6cef1f"} Oct 04 07:25:32 crc kubenswrapper[4987]: I1004 07:25:32.187154 4987 generic.go:334] "Generic (PLEG): container finished" podID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerID="1c2a9bff2bfff859a1a34b6c5a95636ec471fed26d243ddc1df4e356bf67128f" exitCode=0 Oct 04 07:25:32 crc kubenswrapper[4987]: I1004 07:25:32.187201 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" event={"ID":"82cd04d6-e9cc-4f3a-89f6-d867a256a4af","Type":"ContainerDied","Data":"1c2a9bff2bfff859a1a34b6c5a95636ec471fed26d243ddc1df4e356bf67128f"} Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.439777 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.552618 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tscjx\" (UniqueName: \"kubernetes.io/projected/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-kube-api-access-tscjx\") pod \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.552765 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-util\") pod \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.552834 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-bundle\") pod \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\" (UID: \"82cd04d6-e9cc-4f3a-89f6-d867a256a4af\") " Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.554091 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-bundle" (OuterVolumeSpecName: "bundle") pod "82cd04d6-e9cc-4f3a-89f6-d867a256a4af" (UID: "82cd04d6-e9cc-4f3a-89f6-d867a256a4af"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.559617 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-kube-api-access-tscjx" (OuterVolumeSpecName: "kube-api-access-tscjx") pod "82cd04d6-e9cc-4f3a-89f6-d867a256a4af" (UID: "82cd04d6-e9cc-4f3a-89f6-d867a256a4af"). InnerVolumeSpecName "kube-api-access-tscjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.568133 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-util" (OuterVolumeSpecName: "util") pod "82cd04d6-e9cc-4f3a-89f6-d867a256a4af" (UID: "82cd04d6-e9cc-4f3a-89f6-d867a256a4af"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.655085 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.655127 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:25:33 crc kubenswrapper[4987]: I1004 07:25:33.655142 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tscjx\" (UniqueName: \"kubernetes.io/projected/82cd04d6-e9cc-4f3a-89f6-d867a256a4af-kube-api-access-tscjx\") on node \"crc\" DevicePath \"\"" Oct 04 07:25:34 crc kubenswrapper[4987]: I1004 07:25:34.201519 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" event={"ID":"82cd04d6-e9cc-4f3a-89f6-d867a256a4af","Type":"ContainerDied","Data":"76acf1f995c1ca499b6356d67c5fd48726bdfbcad6d3dd1fcada0cd3b429d9ff"} Oct 04 07:25:34 crc kubenswrapper[4987]: I1004 07:25:34.201567 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76acf1f995c1ca499b6356d67c5fd48726bdfbcad6d3dd1fcada0cd3b429d9ff" Oct 04 07:25:34 crc kubenswrapper[4987]: I1004 07:25:34.201652 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.130968 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx"] Oct 04 07:25:39 crc kubenswrapper[4987]: E1004 07:25:39.131742 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="util" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.131755 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="util" Oct 04 07:25:39 crc kubenswrapper[4987]: E1004 07:25:39.131764 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="extract" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.131770 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="extract" Oct 04 07:25:39 crc kubenswrapper[4987]: E1004 07:25:39.131784 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="pull" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.131790 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="pull" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.131925 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="82cd04d6-e9cc-4f3a-89f6-d867a256a4af" containerName="extract" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.132392 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.145400 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-zjcrq" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.185092 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx"] Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.236014 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb65j\" (UniqueName: \"kubernetes.io/projected/d3f2f50a-bd5f-4182-af2f-edd6bce09875-kube-api-access-jb65j\") pod \"rabbitmq-cluster-operator-779fc9694b-rvmfx\" (UID: \"d3f2f50a-bd5f-4182-af2f-edd6bce09875\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.336973 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb65j\" (UniqueName: \"kubernetes.io/projected/d3f2f50a-bd5f-4182-af2f-edd6bce09875-kube-api-access-jb65j\") pod \"rabbitmq-cluster-operator-779fc9694b-rvmfx\" (UID: \"d3f2f50a-bd5f-4182-af2f-edd6bce09875\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.357439 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb65j\" (UniqueName: \"kubernetes.io/projected/d3f2f50a-bd5f-4182-af2f-edd6bce09875-kube-api-access-jb65j\") pod \"rabbitmq-cluster-operator-779fc9694b-rvmfx\" (UID: \"d3f2f50a-bd5f-4182-af2f-edd6bce09875\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.450489 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" Oct 04 07:25:39 crc kubenswrapper[4987]: I1004 07:25:39.677532 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx"] Oct 04 07:25:40 crc kubenswrapper[4987]: I1004 07:25:40.242328 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" event={"ID":"d3f2f50a-bd5f-4182-af2f-edd6bce09875","Type":"ContainerStarted","Data":"0711e6e6009bec711ccb206e2529b36e00b0980df2eaefaebb56d71a8d4ad06f"} Oct 04 07:25:42 crc kubenswrapper[4987]: I1004 07:25:42.256049 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" event={"ID":"d3f2f50a-bd5f-4182-af2f-edd6bce09875","Type":"ContainerStarted","Data":"c000232089ba5e75d1006e1788e6e5fe37ccd66d11676d656dbd69ca97b1ff20"} Oct 04 07:25:42 crc kubenswrapper[4987]: I1004 07:25:42.273222 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-rvmfx" podStartSLOduration=1.447938307 podStartE2EDuration="3.273200051s" podCreationTimestamp="2025-10-04 07:25:39 +0000 UTC" firstStartedPulling="2025-10-04 07:25:39.689114868 +0000 UTC m=+956.744013710" lastFinishedPulling="2025-10-04 07:25:41.514376612 +0000 UTC m=+958.569275454" observedRunningTime="2025-10-04 07:25:42.268063179 +0000 UTC m=+959.322962021" watchObservedRunningTime="2025-10-04 07:25:42.273200051 +0000 UTC m=+959.328098893" Oct 04 07:25:42 crc kubenswrapper[4987]: I1004 07:25:42.703208 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:25:42 crc kubenswrapper[4987]: I1004 07:25:42.703580 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.595776 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.597403 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.599325 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.599974 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.600959 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.601230 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.605470 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-qx7jf" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.611386 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758197 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758294 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bcde8202-83ce-4dfa-9562-4733938935f0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758320 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758487 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4e86e867-43f4-4402-a79f-c04511f29d20\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e86e867-43f4-4402-a79f-c04511f29d20\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758550 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bcde8202-83ce-4dfa-9562-4733938935f0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758704 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs6kv\" (UniqueName: \"kubernetes.io/projected/bcde8202-83ce-4dfa-9562-4733938935f0-kube-api-access-fs6kv\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758765 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.758828 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bcde8202-83ce-4dfa-9562-4733938935f0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860580 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860674 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bcde8202-83ce-4dfa-9562-4733938935f0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860703 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860762 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4e86e867-43f4-4402-a79f-c04511f29d20\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e86e867-43f4-4402-a79f-c04511f29d20\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860787 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bcde8202-83ce-4dfa-9562-4733938935f0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860816 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs6kv\" (UniqueName: \"kubernetes.io/projected/bcde8202-83ce-4dfa-9562-4733938935f0-kube-api-access-fs6kv\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860835 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.860856 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bcde8202-83ce-4dfa-9562-4733938935f0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.861773 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.862082 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.862542 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bcde8202-83ce-4dfa-9562-4733938935f0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.866698 4987 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.866750 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4e86e867-43f4-4402-a79f-c04511f29d20\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e86e867-43f4-4402-a79f-c04511f29d20\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/edf60c18d164d15031d888ea271606e744db3e67b99ca93f74478f6596754d5c/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.867244 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bcde8202-83ce-4dfa-9562-4733938935f0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.867261 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bcde8202-83ce-4dfa-9562-4733938935f0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.870161 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bcde8202-83ce-4dfa-9562-4733938935f0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.886722 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs6kv\" (UniqueName: \"kubernetes.io/projected/bcde8202-83ce-4dfa-9562-4733938935f0-kube-api-access-fs6kv\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.897293 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4e86e867-43f4-4402-a79f-c04511f29d20\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e86e867-43f4-4402-a79f-c04511f29d20\") pod \"rabbitmq-server-0\" (UID: \"bcde8202-83ce-4dfa-9562-4733938935f0\") " pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:47 crc kubenswrapper[4987]: I1004 07:25:47.919516 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:25:48 crc kubenswrapper[4987]: I1004 07:25:48.334848 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Oct 04 07:25:48 crc kubenswrapper[4987]: W1004 07:25:48.341165 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcde8202_83ce_4dfa_9562_4733938935f0.slice/crio-14e8044c22d2bb329e761532eafc4278257817a9d52cc9bb1c447b28e00abc9d WatchSource:0}: Error finding container 14e8044c22d2bb329e761532eafc4278257817a9d52cc9bb1c447b28e00abc9d: Status 404 returned error can't find the container with id 14e8044c22d2bb329e761532eafc4278257817a9d52cc9bb1c447b28e00abc9d Oct 04 07:25:49 crc kubenswrapper[4987]: I1004 07:25:49.300222 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bcde8202-83ce-4dfa-9562-4733938935f0","Type":"ContainerStarted","Data":"14e8044c22d2bb329e761532eafc4278257817a9d52cc9bb1c447b28e00abc9d"} Oct 04 07:25:54 crc kubenswrapper[4987]: I1004 07:25:54.331404 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bcde8202-83ce-4dfa-9562-4733938935f0","Type":"ContainerStarted","Data":"651161b7bd6530c665920a60887e13112b495629574c511b0db038e25cddd995"} Oct 04 07:25:58 crc kubenswrapper[4987]: I1004 07:25:58.901009 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-bcsxf"] Oct 04 07:25:58 crc kubenswrapper[4987]: I1004 07:25:58.902321 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:25:58 crc kubenswrapper[4987]: I1004 07:25:58.904426 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-lff4j" Oct 04 07:25:58 crc kubenswrapper[4987]: I1004 07:25:58.911941 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-bcsxf"] Oct 04 07:25:59 crc kubenswrapper[4987]: I1004 07:25:59.065673 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqkrh\" (UniqueName: \"kubernetes.io/projected/1d66401a-3587-4fdd-9a4c-fac02692709d-kube-api-access-sqkrh\") pod \"keystone-operator-index-bcsxf\" (UID: \"1d66401a-3587-4fdd-9a4c-fac02692709d\") " pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:25:59 crc kubenswrapper[4987]: I1004 07:25:59.166935 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqkrh\" (UniqueName: \"kubernetes.io/projected/1d66401a-3587-4fdd-9a4c-fac02692709d-kube-api-access-sqkrh\") pod \"keystone-operator-index-bcsxf\" (UID: \"1d66401a-3587-4fdd-9a4c-fac02692709d\") " pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:25:59 crc kubenswrapper[4987]: I1004 07:25:59.186442 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqkrh\" (UniqueName: \"kubernetes.io/projected/1d66401a-3587-4fdd-9a4c-fac02692709d-kube-api-access-sqkrh\") pod \"keystone-operator-index-bcsxf\" (UID: \"1d66401a-3587-4fdd-9a4c-fac02692709d\") " pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:25:59 crc kubenswrapper[4987]: I1004 07:25:59.219322 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:25:59 crc kubenswrapper[4987]: I1004 07:25:59.408173 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-bcsxf"] Oct 04 07:25:59 crc kubenswrapper[4987]: W1004 07:25:59.411443 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d66401a_3587_4fdd_9a4c_fac02692709d.slice/crio-c72df5cd70ada77b6bb5c8a31786720213e6684c2c3d2f39ccfe10e90c349983 WatchSource:0}: Error finding container c72df5cd70ada77b6bb5c8a31786720213e6684c2c3d2f39ccfe10e90c349983: Status 404 returned error can't find the container with id c72df5cd70ada77b6bb5c8a31786720213e6684c2c3d2f39ccfe10e90c349983 Oct 04 07:26:00 crc kubenswrapper[4987]: I1004 07:26:00.368879 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-bcsxf" event={"ID":"1d66401a-3587-4fdd-9a4c-fac02692709d","Type":"ContainerStarted","Data":"c72df5cd70ada77b6bb5c8a31786720213e6684c2c3d2f39ccfe10e90c349983"} Oct 04 07:26:01 crc kubenswrapper[4987]: I1004 07:26:01.385347 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-bcsxf" event={"ID":"1d66401a-3587-4fdd-9a4c-fac02692709d","Type":"ContainerStarted","Data":"72ecbe2cc98b3d0639196d4f899a5947d2781596ce4e30589b2b2e2b958b42ca"} Oct 04 07:26:01 crc kubenswrapper[4987]: I1004 07:26:01.411703 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-bcsxf" podStartSLOduration=2.430533008 podStartE2EDuration="3.411676542s" podCreationTimestamp="2025-10-04 07:25:58 +0000 UTC" firstStartedPulling="2025-10-04 07:25:59.413784524 +0000 UTC m=+976.468683366" lastFinishedPulling="2025-10-04 07:26:00.394928058 +0000 UTC m=+977.449826900" observedRunningTime="2025-10-04 07:26:01.404160628 +0000 UTC m=+978.459059470" watchObservedRunningTime="2025-10-04 07:26:01.411676542 +0000 UTC m=+978.466575394" Oct 04 07:26:09 crc kubenswrapper[4987]: I1004 07:26:09.219753 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:26:09 crc kubenswrapper[4987]: I1004 07:26:09.220476 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:26:09 crc kubenswrapper[4987]: I1004 07:26:09.246682 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:26:09 crc kubenswrapper[4987]: I1004 07:26:09.486043 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-bcsxf" Oct 04 07:26:12 crc kubenswrapper[4987]: I1004 07:26:12.703221 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:26:12 crc kubenswrapper[4987]: I1004 07:26:12.704027 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:26:12 crc kubenswrapper[4987]: I1004 07:26:12.704083 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:26:12 crc kubenswrapper[4987]: I1004 07:26:12.704822 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"914ae0228b00a348da279a19d0c789018ccba172664058b29703314fa8fd5ab8"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:26:12 crc kubenswrapper[4987]: I1004 07:26:12.704881 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://914ae0228b00a348da279a19d0c789018ccba172664058b29703314fa8fd5ab8" gracePeriod=600 Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.454919 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="914ae0228b00a348da279a19d0c789018ccba172664058b29703314fa8fd5ab8" exitCode=0 Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.454989 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"914ae0228b00a348da279a19d0c789018ccba172664058b29703314fa8fd5ab8"} Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.455411 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"9851ab3918efc181099219aac8c4cb4de265cdf08bb6a6a2096a913ebd0e22d6"} Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.455442 4987 scope.go:117] "RemoveContainer" containerID="241b883c886b42b50a0701b97bca81215e874c5bdbc8a5d84d217f37244c27ff" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.542826 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk"] Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.544325 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.546834 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwsts" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.566211 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk"] Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.575028 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-util\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.575413 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg7k6\" (UniqueName: \"kubernetes.io/projected/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-kube-api-access-vg7k6\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.575774 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-bundle\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.676762 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-bundle\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.676838 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-util\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.676886 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg7k6\" (UniqueName: \"kubernetes.io/projected/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-kube-api-access-vg7k6\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.677712 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-bundle\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.677760 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-util\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.695285 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg7k6\" (UniqueName: \"kubernetes.io/projected/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-kube-api-access-vg7k6\") pod \"8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:13 crc kubenswrapper[4987]: I1004 07:26:13.860135 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:14 crc kubenswrapper[4987]: I1004 07:26:14.247852 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk"] Oct 04 07:26:14 crc kubenswrapper[4987]: W1004 07:26:14.256029 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32ec6fc4_2f39_4e20_b529_9aa82a827fb7.slice/crio-ba29fee4e26ba67b5133a93aa68bc3a006a997bdfdf6c36efd490d24baf8897d WatchSource:0}: Error finding container ba29fee4e26ba67b5133a93aa68bc3a006a997bdfdf6c36efd490d24baf8897d: Status 404 returned error can't find the container with id ba29fee4e26ba67b5133a93aa68bc3a006a997bdfdf6c36efd490d24baf8897d Oct 04 07:26:14 crc kubenswrapper[4987]: I1004 07:26:14.478336 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerStarted","Data":"972cfd2e0e2017637b571070fbd3b6eb5c35b6afe4713e2150f607d838116d86"} Oct 04 07:26:14 crc kubenswrapper[4987]: I1004 07:26:14.478733 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerStarted","Data":"ba29fee4e26ba67b5133a93aa68bc3a006a997bdfdf6c36efd490d24baf8897d"} Oct 04 07:26:15 crc kubenswrapper[4987]: I1004 07:26:15.494501 4987 generic.go:334] "Generic (PLEG): container finished" podID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerID="972cfd2e0e2017637b571070fbd3b6eb5c35b6afe4713e2150f607d838116d86" exitCode=0 Oct 04 07:26:15 crc kubenswrapper[4987]: I1004 07:26:15.494550 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerDied","Data":"972cfd2e0e2017637b571070fbd3b6eb5c35b6afe4713e2150f607d838116d86"} Oct 04 07:26:17 crc kubenswrapper[4987]: I1004 07:26:17.526935 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerStarted","Data":"769f192d84094c5c69b5498fbd55933fb7eaf67cba8b011826cb196bd16ddee5"} Oct 04 07:26:18 crc kubenswrapper[4987]: I1004 07:26:18.535038 4987 generic.go:334] "Generic (PLEG): container finished" podID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerID="769f192d84094c5c69b5498fbd55933fb7eaf67cba8b011826cb196bd16ddee5" exitCode=0 Oct 04 07:26:18 crc kubenswrapper[4987]: I1004 07:26:18.535153 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerDied","Data":"769f192d84094c5c69b5498fbd55933fb7eaf67cba8b011826cb196bd16ddee5"} Oct 04 07:26:19 crc kubenswrapper[4987]: I1004 07:26:19.545576 4987 generic.go:334] "Generic (PLEG): container finished" podID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerID="6ef0546b7dedc7e2c02e1e91ace7f24fb74c5300e436278511930e180ec0b0cb" exitCode=0 Oct 04 07:26:19 crc kubenswrapper[4987]: I1004 07:26:19.545704 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerDied","Data":"6ef0546b7dedc7e2c02e1e91ace7f24fb74c5300e436278511930e180ec0b0cb"} Oct 04 07:26:20 crc kubenswrapper[4987]: I1004 07:26:20.834771 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:20 crc kubenswrapper[4987]: I1004 07:26:20.993945 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg7k6\" (UniqueName: \"kubernetes.io/projected/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-kube-api-access-vg7k6\") pod \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " Oct 04 07:26:20 crc kubenswrapper[4987]: I1004 07:26:20.994133 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-util\") pod \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " Oct 04 07:26:20 crc kubenswrapper[4987]: I1004 07:26:20.994280 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-bundle\") pod \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\" (UID: \"32ec6fc4-2f39-4e20-b529-9aa82a827fb7\") " Oct 04 07:26:20 crc kubenswrapper[4987]: I1004 07:26:20.995572 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-bundle" (OuterVolumeSpecName: "bundle") pod "32ec6fc4-2f39-4e20-b529-9aa82a827fb7" (UID: "32ec6fc4-2f39-4e20-b529-9aa82a827fb7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.003446 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-kube-api-access-vg7k6" (OuterVolumeSpecName: "kube-api-access-vg7k6") pod "32ec6fc4-2f39-4e20-b529-9aa82a827fb7" (UID: "32ec6fc4-2f39-4e20-b529-9aa82a827fb7"). InnerVolumeSpecName "kube-api-access-vg7k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.004076 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-util" (OuterVolumeSpecName: "util") pod "32ec6fc4-2f39-4e20-b529-9aa82a827fb7" (UID: "32ec6fc4-2f39-4e20-b529-9aa82a827fb7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.096943 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.097005 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg7k6\" (UniqueName: \"kubernetes.io/projected/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-kube-api-access-vg7k6\") on node \"crc\" DevicePath \"\"" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.097094 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/32ec6fc4-2f39-4e20-b529-9aa82a827fb7-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.561613 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" event={"ID":"32ec6fc4-2f39-4e20-b529-9aa82a827fb7","Type":"ContainerDied","Data":"ba29fee4e26ba67b5133a93aa68bc3a006a997bdfdf6c36efd490d24baf8897d"} Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.561727 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba29fee4e26ba67b5133a93aa68bc3a006a997bdfdf6c36efd490d24baf8897d" Oct 04 07:26:21 crc kubenswrapper[4987]: I1004 07:26:21.561669 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk" Oct 04 07:26:26 crc kubenswrapper[4987]: I1004 07:26:26.595226 4987 generic.go:334] "Generic (PLEG): container finished" podID="bcde8202-83ce-4dfa-9562-4733938935f0" containerID="651161b7bd6530c665920a60887e13112b495629574c511b0db038e25cddd995" exitCode=0 Oct 04 07:26:26 crc kubenswrapper[4987]: I1004 07:26:26.595265 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bcde8202-83ce-4dfa-9562-4733938935f0","Type":"ContainerDied","Data":"651161b7bd6530c665920a60887e13112b495629574c511b0db038e25cddd995"} Oct 04 07:26:27 crc kubenswrapper[4987]: I1004 07:26:27.604118 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"bcde8202-83ce-4dfa-9562-4733938935f0","Type":"ContainerStarted","Data":"fb057bd61ad1165525a4cae3ad5fd11b584cb0ad09e442c6fede8db52d831924"} Oct 04 07:26:27 crc kubenswrapper[4987]: I1004 07:26:27.604797 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:26:27 crc kubenswrapper[4987]: I1004 07:26:27.627216 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.783757321 podStartE2EDuration="41.627199044s" podCreationTimestamp="2025-10-04 07:25:46 +0000 UTC" firstStartedPulling="2025-10-04 07:25:48.344743797 +0000 UTC m=+965.399642639" lastFinishedPulling="2025-10-04 07:25:53.18818552 +0000 UTC m=+970.243084362" observedRunningTime="2025-10-04 07:26:27.623625383 +0000 UTC m=+1004.678524245" watchObservedRunningTime="2025-10-04 07:26:27.627199044 +0000 UTC m=+1004.682097886" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.702245 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7"] Oct 04 07:26:28 crc kubenswrapper[4987]: E1004 07:26:28.702586 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="util" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.702602 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="util" Oct 04 07:26:28 crc kubenswrapper[4987]: E1004 07:26:28.702615 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="extract" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.702624 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="extract" Oct 04 07:26:28 crc kubenswrapper[4987]: E1004 07:26:28.702642 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="pull" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.702674 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="pull" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.702875 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ec6fc4-2f39-4e20-b529-9aa82a827fb7" containerName="extract" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.703870 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.706480 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.707740 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-c7b9l" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.735533 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7"] Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.819059 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-webhook-cert\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.819121 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-apiservice-cert\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.819160 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42fzz\" (UniqueName: \"kubernetes.io/projected/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-kube-api-access-42fzz\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.921361 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-webhook-cert\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.921475 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-apiservice-cert\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.921569 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42fzz\" (UniqueName: \"kubernetes.io/projected/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-kube-api-access-42fzz\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.933423 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-apiservice-cert\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.933481 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-webhook-cert\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:28 crc kubenswrapper[4987]: I1004 07:26:28.949478 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42fzz\" (UniqueName: \"kubernetes.io/projected/81bd06d1-6dc2-4132-9e01-2d3b72a69cea-kube-api-access-42fzz\") pod \"keystone-operator-controller-manager-55fd955c48-qdph7\" (UID: \"81bd06d1-6dc2-4132-9e01-2d3b72a69cea\") " pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:29 crc kubenswrapper[4987]: I1004 07:26:29.024744 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:29 crc kubenswrapper[4987]: I1004 07:26:29.486344 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7"] Oct 04 07:26:29 crc kubenswrapper[4987]: W1004 07:26:29.489943 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81bd06d1_6dc2_4132_9e01_2d3b72a69cea.slice/crio-cb2b3edd5cd40af896a7c119f35e3db35cef73ad1cd38d933f97e6e7435d017f WatchSource:0}: Error finding container cb2b3edd5cd40af896a7c119f35e3db35cef73ad1cd38d933f97e6e7435d017f: Status 404 returned error can't find the container with id cb2b3edd5cd40af896a7c119f35e3db35cef73ad1cd38d933f97e6e7435d017f Oct 04 07:26:29 crc kubenswrapper[4987]: I1004 07:26:29.626300 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" event={"ID":"81bd06d1-6dc2-4132-9e01-2d3b72a69cea","Type":"ContainerStarted","Data":"cb2b3edd5cd40af896a7c119f35e3db35cef73ad1cd38d933f97e6e7435d017f"} Oct 04 07:26:32 crc kubenswrapper[4987]: I1004 07:26:32.653479 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" event={"ID":"81bd06d1-6dc2-4132-9e01-2d3b72a69cea","Type":"ContainerStarted","Data":"a5d9358ebcc36c5e39c6ca8523ffbd30c3fae22eede240f7e6055c0a76b6a871"} Oct 04 07:26:33 crc kubenswrapper[4987]: I1004 07:26:33.662356 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" event={"ID":"81bd06d1-6dc2-4132-9e01-2d3b72a69cea","Type":"ContainerStarted","Data":"dc49adcf68e46c1011ec6d355938dd379cf218e23d251b3fe10612e0133e7af4"} Oct 04 07:26:33 crc kubenswrapper[4987]: I1004 07:26:33.662927 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:33 crc kubenswrapper[4987]: I1004 07:26:33.684222 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" podStartSLOduration=3.175949395 podStartE2EDuration="5.684195162s" podCreationTimestamp="2025-10-04 07:26:28 +0000 UTC" firstStartedPulling="2025-10-04 07:26:29.492982661 +0000 UTC m=+1006.547881503" lastFinishedPulling="2025-10-04 07:26:32.001228428 +0000 UTC m=+1009.056127270" observedRunningTime="2025-10-04 07:26:33.682862581 +0000 UTC m=+1010.737761433" watchObservedRunningTime="2025-10-04 07:26:33.684195162 +0000 UTC m=+1010.739094004" Oct 04 07:26:37 crc kubenswrapper[4987]: I1004 07:26:37.923755 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Oct 04 07:26:39 crc kubenswrapper[4987]: I1004 07:26:39.029294 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55fd955c48-qdph7" Oct 04 07:26:41 crc kubenswrapper[4987]: I1004 07:26:41.777712 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-cww6m"] Oct 04 07:26:41 crc kubenswrapper[4987]: I1004 07:26:41.780438 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:41 crc kubenswrapper[4987]: I1004 07:26:41.825849 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-cww6m"] Oct 04 07:26:41 crc kubenswrapper[4987]: I1004 07:26:41.942532 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhz7k\" (UniqueName: \"kubernetes.io/projected/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2-kube-api-access-fhz7k\") pod \"keystone-db-create-cww6m\" (UID: \"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2\") " pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:42 crc kubenswrapper[4987]: I1004 07:26:42.044046 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhz7k\" (UniqueName: \"kubernetes.io/projected/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2-kube-api-access-fhz7k\") pod \"keystone-db-create-cww6m\" (UID: \"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2\") " pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:42 crc kubenswrapper[4987]: I1004 07:26:42.067733 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhz7k\" (UniqueName: \"kubernetes.io/projected/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2-kube-api-access-fhz7k\") pod \"keystone-db-create-cww6m\" (UID: \"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2\") " pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:42 crc kubenswrapper[4987]: I1004 07:26:42.150416 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:42 crc kubenswrapper[4987]: I1004 07:26:42.566136 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-cww6m"] Oct 04 07:26:42 crc kubenswrapper[4987]: I1004 07:26:42.719019 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-cww6m" event={"ID":"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2","Type":"ContainerStarted","Data":"694b5ea22e509b6a3e6b8c6dea14bd5700292f87de926a44778033a5e7f39fbf"} Oct 04 07:26:43 crc kubenswrapper[4987]: I1004 07:26:43.729985 4987 generic.go:334] "Generic (PLEG): container finished" podID="a93fd53a-dad1-4a5a-8e9e-50c393cac9b2" containerID="b380672c2863904357879ef19930b152e78463e79fe93dfe15304bdd7fc36157" exitCode=0 Oct 04 07:26:43 crc kubenswrapper[4987]: I1004 07:26:43.730040 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-cww6m" event={"ID":"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2","Type":"ContainerDied","Data":"b380672c2863904357879ef19930b152e78463e79fe93dfe15304bdd7fc36157"} Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.033792 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.199560 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhz7k\" (UniqueName: \"kubernetes.io/projected/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2-kube-api-access-fhz7k\") pod \"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2\" (UID: \"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2\") " Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.205777 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2-kube-api-access-fhz7k" (OuterVolumeSpecName: "kube-api-access-fhz7k") pod "a93fd53a-dad1-4a5a-8e9e-50c393cac9b2" (UID: "a93fd53a-dad1-4a5a-8e9e-50c393cac9b2"). InnerVolumeSpecName "kube-api-access-fhz7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.301567 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhz7k\" (UniqueName: \"kubernetes.io/projected/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2-kube-api-access-fhz7k\") on node \"crc\" DevicePath \"\"" Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.753885 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-cww6m" event={"ID":"a93fd53a-dad1-4a5a-8e9e-50c393cac9b2","Type":"ContainerDied","Data":"694b5ea22e509b6a3e6b8c6dea14bd5700292f87de926a44778033a5e7f39fbf"} Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.753934 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-cww6m" Oct 04 07:26:45 crc kubenswrapper[4987]: I1004 07:26:45.753947 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="694b5ea22e509b6a3e6b8c6dea14bd5700292f87de926a44778033a5e7f39fbf" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.693324 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-c0c2-account-create-blbdj"] Oct 04 07:26:51 crc kubenswrapper[4987]: E1004 07:26:51.694026 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a93fd53a-dad1-4a5a-8e9e-50c393cac9b2" containerName="mariadb-database-create" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.694045 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a93fd53a-dad1-4a5a-8e9e-50c393cac9b2" containerName="mariadb-database-create" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.694189 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="a93fd53a-dad1-4a5a-8e9e-50c393cac9b2" containerName="mariadb-database-create" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.694755 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.697509 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.717909 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-c0c2-account-create-blbdj"] Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.799231 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls452\" (UniqueName: \"kubernetes.io/projected/9eff36d2-e701-4996-b283-cfbbc1b27e0c-kube-api-access-ls452\") pod \"keystone-c0c2-account-create-blbdj\" (UID: \"9eff36d2-e701-4996-b283-cfbbc1b27e0c\") " pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.900869 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls452\" (UniqueName: \"kubernetes.io/projected/9eff36d2-e701-4996-b283-cfbbc1b27e0c-kube-api-access-ls452\") pod \"keystone-c0c2-account-create-blbdj\" (UID: \"9eff36d2-e701-4996-b283-cfbbc1b27e0c\") " pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:51 crc kubenswrapper[4987]: I1004 07:26:51.923556 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls452\" (UniqueName: \"kubernetes.io/projected/9eff36d2-e701-4996-b283-cfbbc1b27e0c-kube-api-access-ls452\") pod \"keystone-c0c2-account-create-blbdj\" (UID: \"9eff36d2-e701-4996-b283-cfbbc1b27e0c\") " pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:52 crc kubenswrapper[4987]: I1004 07:26:52.013975 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:52 crc kubenswrapper[4987]: I1004 07:26:52.429949 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-c0c2-account-create-blbdj"] Oct 04 07:26:52 crc kubenswrapper[4987]: I1004 07:26:52.805339 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" event={"ID":"9eff36d2-e701-4996-b283-cfbbc1b27e0c","Type":"ContainerStarted","Data":"14a1def2b6c691fa8461c6ab4d716c08f3b62a75097feef3c4dd69676fee4592"} Oct 04 07:26:52 crc kubenswrapper[4987]: I1004 07:26:52.805448 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" event={"ID":"9eff36d2-e701-4996-b283-cfbbc1b27e0c","Type":"ContainerStarted","Data":"9d09bbb9dfd08bff279055ca026a521619e18183477cb6fd60e47a815ed0470c"} Oct 04 07:26:52 crc kubenswrapper[4987]: I1004 07:26:52.830748 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" podStartSLOduration=1.830729656 podStartE2EDuration="1.830729656s" podCreationTimestamp="2025-10-04 07:26:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:26:52.825898936 +0000 UTC m=+1029.880797818" watchObservedRunningTime="2025-10-04 07:26:52.830729656 +0000 UTC m=+1029.885628498" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.106775 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-2kdfj"] Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.107903 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.110608 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-v4dck" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.131837 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-2kdfj"] Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.223318 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskmt\" (UniqueName: \"kubernetes.io/projected/09609231-954a-4b7a-9878-e4f618347a47-kube-api-access-gskmt\") pod \"horizon-operator-index-2kdfj\" (UID: \"09609231-954a-4b7a-9878-e4f618347a47\") " pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.324703 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskmt\" (UniqueName: \"kubernetes.io/projected/09609231-954a-4b7a-9878-e4f618347a47-kube-api-access-gskmt\") pod \"horizon-operator-index-2kdfj\" (UID: \"09609231-954a-4b7a-9878-e4f618347a47\") " pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.353784 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskmt\" (UniqueName: \"kubernetes.io/projected/09609231-954a-4b7a-9878-e4f618347a47-kube-api-access-gskmt\") pod \"horizon-operator-index-2kdfj\" (UID: \"09609231-954a-4b7a-9878-e4f618347a47\") " pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.422801 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.825891 4987 generic.go:334] "Generic (PLEG): container finished" podID="9eff36d2-e701-4996-b283-cfbbc1b27e0c" containerID="14a1def2b6c691fa8461c6ab4d716c08f3b62a75097feef3c4dd69676fee4592" exitCode=0 Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.826916 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" event={"ID":"9eff36d2-e701-4996-b283-cfbbc1b27e0c","Type":"ContainerDied","Data":"14a1def2b6c691fa8461c6ab4d716c08f3b62a75097feef3c4dd69676fee4592"} Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.885072 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-2kdfj"] Oct 04 07:26:53 crc kubenswrapper[4987]: W1004 07:26:53.900370 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09609231_954a_4b7a_9878_e4f618347a47.slice/crio-7eaa85cf563e5caa43a22d750a17b80682917f34b260796ed36327557141f2f7 WatchSource:0}: Error finding container 7eaa85cf563e5caa43a22d750a17b80682917f34b260796ed36327557141f2f7: Status 404 returned error can't find the container with id 7eaa85cf563e5caa43a22d750a17b80682917f34b260796ed36327557141f2f7 Oct 04 07:26:53 crc kubenswrapper[4987]: I1004 07:26:53.903936 4987 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.500044 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-qjl2t"] Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.501523 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.504396 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-hvs4z" Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.516379 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-qjl2t"] Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.562916 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzxq8\" (UniqueName: \"kubernetes.io/projected/73a59a02-da82-4edc-bded-8d2fc1484bfd-kube-api-access-qzxq8\") pod \"swift-operator-index-qjl2t\" (UID: \"73a59a02-da82-4edc-bded-8d2fc1484bfd\") " pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.664244 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzxq8\" (UniqueName: \"kubernetes.io/projected/73a59a02-da82-4edc-bded-8d2fc1484bfd-kube-api-access-qzxq8\") pod \"swift-operator-index-qjl2t\" (UID: \"73a59a02-da82-4edc-bded-8d2fc1484bfd\") " pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.687732 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzxq8\" (UniqueName: \"kubernetes.io/projected/73a59a02-da82-4edc-bded-8d2fc1484bfd-kube-api-access-qzxq8\") pod \"swift-operator-index-qjl2t\" (UID: \"73a59a02-da82-4edc-bded-8d2fc1484bfd\") " pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.833644 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:26:54 crc kubenswrapper[4987]: I1004 07:26:54.844377 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-2kdfj" event={"ID":"09609231-954a-4b7a-9878-e4f618347a47","Type":"ContainerStarted","Data":"7eaa85cf563e5caa43a22d750a17b80682917f34b260796ed36327557141f2f7"} Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.291360 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-qjl2t"] Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.298157 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.482700 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls452\" (UniqueName: \"kubernetes.io/projected/9eff36d2-e701-4996-b283-cfbbc1b27e0c-kube-api-access-ls452\") pod \"9eff36d2-e701-4996-b283-cfbbc1b27e0c\" (UID: \"9eff36d2-e701-4996-b283-cfbbc1b27e0c\") " Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.490812 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eff36d2-e701-4996-b283-cfbbc1b27e0c-kube-api-access-ls452" (OuterVolumeSpecName: "kube-api-access-ls452") pod "9eff36d2-e701-4996-b283-cfbbc1b27e0c" (UID: "9eff36d2-e701-4996-b283-cfbbc1b27e0c"). InnerVolumeSpecName "kube-api-access-ls452". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.586066 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls452\" (UniqueName: \"kubernetes.io/projected/9eff36d2-e701-4996-b283-cfbbc1b27e0c-kube-api-access-ls452\") on node \"crc\" DevicePath \"\"" Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.852421 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-qjl2t" event={"ID":"73a59a02-da82-4edc-bded-8d2fc1484bfd","Type":"ContainerStarted","Data":"976f013299352159b57297b6840fcdaa882796786c1f77813239346f0b93a6c3"} Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.854830 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-2kdfj" event={"ID":"09609231-954a-4b7a-9878-e4f618347a47","Type":"ContainerStarted","Data":"eb4d0d4d6561dacc11f3a534d0e61be81b7ac353b4bd70e90fe15e47420b8c05"} Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.856876 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" event={"ID":"9eff36d2-e701-4996-b283-cfbbc1b27e0c","Type":"ContainerDied","Data":"9d09bbb9dfd08bff279055ca026a521619e18183477cb6fd60e47a815ed0470c"} Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.856903 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d09bbb9dfd08bff279055ca026a521619e18183477cb6fd60e47a815ed0470c" Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.856944 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-c0c2-account-create-blbdj" Oct 04 07:26:55 crc kubenswrapper[4987]: I1004 07:26:55.873795 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-2kdfj" podStartSLOduration=1.356370284 podStartE2EDuration="2.873771523s" podCreationTimestamp="2025-10-04 07:26:53 +0000 UTC" firstStartedPulling="2025-10-04 07:26:53.903558467 +0000 UTC m=+1030.958457309" lastFinishedPulling="2025-10-04 07:26:55.420959706 +0000 UTC m=+1032.475858548" observedRunningTime="2025-10-04 07:26:55.871050508 +0000 UTC m=+1032.925949340" watchObservedRunningTime="2025-10-04 07:26:55.873771523 +0000 UTC m=+1032.928670365" Oct 04 07:26:56 crc kubenswrapper[4987]: I1004 07:26:56.867427 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-qjl2t" event={"ID":"73a59a02-da82-4edc-bded-8d2fc1484bfd","Type":"ContainerStarted","Data":"56d805b3a209650e42f9b2e1255dcf0f706b4ef02c29ed9c6f22c5f9ebfbbce9"} Oct 04 07:26:56 crc kubenswrapper[4987]: I1004 07:26:56.887613 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-qjl2t" podStartSLOduration=1.9448965280000001 podStartE2EDuration="2.887586406s" podCreationTimestamp="2025-10-04 07:26:54 +0000 UTC" firstStartedPulling="2025-10-04 07:26:55.337085705 +0000 UTC m=+1032.391984547" lastFinishedPulling="2025-10-04 07:26:56.279775593 +0000 UTC m=+1033.334674425" observedRunningTime="2025-10-04 07:26:56.883993724 +0000 UTC m=+1033.938892566" watchObservedRunningTime="2025-10-04 07:26:56.887586406 +0000 UTC m=+1033.942485248" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.266538 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-6sb9w"] Oct 04 07:26:57 crc kubenswrapper[4987]: E1004 07:26:57.266811 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eff36d2-e701-4996-b283-cfbbc1b27e0c" containerName="mariadb-account-create" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.266824 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eff36d2-e701-4996-b283-cfbbc1b27e0c" containerName="mariadb-account-create" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.266948 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eff36d2-e701-4996-b283-cfbbc1b27e0c" containerName="mariadb-account-create" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.267387 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.269665 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-l82gw" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.270168 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.270408 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.276221 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.279371 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-6sb9w"] Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.328715 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2gxf\" (UniqueName: \"kubernetes.io/projected/91962d41-2412-432a-823e-a4351de82e52-kube-api-access-k2gxf\") pod \"keystone-db-sync-6sb9w\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.328954 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91962d41-2412-432a-823e-a4351de82e52-config-data\") pod \"keystone-db-sync-6sb9w\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.430796 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2gxf\" (UniqueName: \"kubernetes.io/projected/91962d41-2412-432a-823e-a4351de82e52-kube-api-access-k2gxf\") pod \"keystone-db-sync-6sb9w\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.430923 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91962d41-2412-432a-823e-a4351de82e52-config-data\") pod \"keystone-db-sync-6sb9w\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.436308 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91962d41-2412-432a-823e-a4351de82e52-config-data\") pod \"keystone-db-sync-6sb9w\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.453949 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2gxf\" (UniqueName: \"kubernetes.io/projected/91962d41-2412-432a-823e-a4351de82e52-kube-api-access-k2gxf\") pod \"keystone-db-sync-6sb9w\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.587912 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.814124 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-6sb9w"] Oct 04 07:26:57 crc kubenswrapper[4987]: W1004 07:26:57.820241 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91962d41_2412_432a_823e_a4351de82e52.slice/crio-c67b22cb5e8ca3a2293e8eaa88fe7090101b06ef92cf835c99a435e831f76fdc WatchSource:0}: Error finding container c67b22cb5e8ca3a2293e8eaa88fe7090101b06ef92cf835c99a435e831f76fdc: Status 404 returned error can't find the container with id c67b22cb5e8ca3a2293e8eaa88fe7090101b06ef92cf835c99a435e831f76fdc Oct 04 07:26:57 crc kubenswrapper[4987]: I1004 07:26:57.876837 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" event={"ID":"91962d41-2412-432a-823e-a4351de82e52","Type":"ContainerStarted","Data":"c67b22cb5e8ca3a2293e8eaa88fe7090101b06ef92cf835c99a435e831f76fdc"} Oct 04 07:27:03 crc kubenswrapper[4987]: I1004 07:27:03.423991 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:27:03 crc kubenswrapper[4987]: I1004 07:27:03.425014 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:27:03 crc kubenswrapper[4987]: I1004 07:27:03.506181 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:27:03 crc kubenswrapper[4987]: I1004 07:27:03.976150 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-2kdfj" Oct 04 07:27:04 crc kubenswrapper[4987]: I1004 07:27:04.835861 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:27:04 crc kubenswrapper[4987]: I1004 07:27:04.837580 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:27:04 crc kubenswrapper[4987]: I1004 07:27:04.870287 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:27:04 crc kubenswrapper[4987]: I1004 07:27:04.988315 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-qjl2t" Oct 04 07:27:06 crc kubenswrapper[4987]: I1004 07:27:06.970479 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" event={"ID":"91962d41-2412-432a-823e-a4351de82e52","Type":"ContainerStarted","Data":"f50f5cdd0a355c36be1e6022a1de82292de621a28d536a0f7bb0735e03d1e5be"} Oct 04 07:27:06 crc kubenswrapper[4987]: I1004 07:27:06.987583 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" podStartSLOduration=1.927812582 podStartE2EDuration="9.987555049s" podCreationTimestamp="2025-10-04 07:26:57 +0000 UTC" firstStartedPulling="2025-10-04 07:26:57.822286105 +0000 UTC m=+1034.877184947" lastFinishedPulling="2025-10-04 07:27:05.882028562 +0000 UTC m=+1042.936927414" observedRunningTime="2025-10-04 07:27:06.986202578 +0000 UTC m=+1044.041101420" watchObservedRunningTime="2025-10-04 07:27:06.987555049 +0000 UTC m=+1044.042453891" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.361150 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd"] Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.365027 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.375889 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwsts" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.395990 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd"] Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.537648 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhd8v\" (UniqueName: \"kubernetes.io/projected/70c28efa-16a5-4eef-8fec-f80bad6cea42-kube-api-access-mhd8v\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.537721 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-util\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.537758 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-bundle\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.639730 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-bundle\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.640043 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhd8v\" (UniqueName: \"kubernetes.io/projected/70c28efa-16a5-4eef-8fec-f80bad6cea42-kube-api-access-mhd8v\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.640104 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-util\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.640741 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-bundle\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.641067 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-util\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.665538 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhd8v\" (UniqueName: \"kubernetes.io/projected/70c28efa-16a5-4eef-8fec-f80bad6cea42-kube-api-access-mhd8v\") pod \"c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.692347 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:10 crc kubenswrapper[4987]: I1004 07:27:10.955458 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd"] Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.004865 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" event={"ID":"70c28efa-16a5-4eef-8fec-f80bad6cea42","Type":"ContainerStarted","Data":"4971f943a12341787c681feb1795ecf731da9265738bf5fc36e52d4439a7a7fd"} Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.146820 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw"] Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.148765 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.159823 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw"] Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.249872 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-util\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.249943 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-bundle\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.249986 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjn5m\" (UniqueName: \"kubernetes.io/projected/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-kube-api-access-qjn5m\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.351164 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-util\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.351218 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-bundle\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.351275 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjn5m\" (UniqueName: \"kubernetes.io/projected/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-kube-api-access-qjn5m\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.351925 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-util\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.352050 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-bundle\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.373917 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjn5m\" (UniqueName: \"kubernetes.io/projected/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-kube-api-access-qjn5m\") pod \"be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:11 crc kubenswrapper[4987]: I1004 07:27:11.474257 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:12 crc kubenswrapper[4987]: I1004 07:27:12.017513 4987 generic.go:334] "Generic (PLEG): container finished" podID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerID="459add402c2c758a9f9d1c48fa0f9a10fafa22d99042b61d3c9e7af4a348cba5" exitCode=0 Oct 04 07:27:12 crc kubenswrapper[4987]: I1004 07:27:12.017602 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" event={"ID":"70c28efa-16a5-4eef-8fec-f80bad6cea42","Type":"ContainerDied","Data":"459add402c2c758a9f9d1c48fa0f9a10fafa22d99042b61d3c9e7af4a348cba5"} Oct 04 07:27:12 crc kubenswrapper[4987]: I1004 07:27:12.182822 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw"] Oct 04 07:27:12 crc kubenswrapper[4987]: W1004 07:27:12.189660 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5d0cde7_7997_4601_b1e1_f8d55d28f8b3.slice/crio-7492e7fdc77249d0e291fdcf638961f03e9a3d74f6a5964ee307f6c73b14743f WatchSource:0}: Error finding container 7492e7fdc77249d0e291fdcf638961f03e9a3d74f6a5964ee307f6c73b14743f: Status 404 returned error can't find the container with id 7492e7fdc77249d0e291fdcf638961f03e9a3d74f6a5964ee307f6c73b14743f Oct 04 07:27:13 crc kubenswrapper[4987]: I1004 07:27:13.031208 4987 generic.go:334] "Generic (PLEG): container finished" podID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerID="59d67704dcfc02c2a2d5c03491a04b47008b35a8d090b0dffebdc3b3593028a7" exitCode=0 Oct 04 07:27:13 crc kubenswrapper[4987]: I1004 07:27:13.031298 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" event={"ID":"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3","Type":"ContainerDied","Data":"59d67704dcfc02c2a2d5c03491a04b47008b35a8d090b0dffebdc3b3593028a7"} Oct 04 07:27:13 crc kubenswrapper[4987]: I1004 07:27:13.032033 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" event={"ID":"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3","Type":"ContainerStarted","Data":"7492e7fdc77249d0e291fdcf638961f03e9a3d74f6a5964ee307f6c73b14743f"} Oct 04 07:27:14 crc kubenswrapper[4987]: I1004 07:27:14.047656 4987 generic.go:334] "Generic (PLEG): container finished" podID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerID="d8d0eed68c5cb89f47a62782bbc3fc4a689f678386e9193651221899eccd6ee5" exitCode=0 Oct 04 07:27:14 crc kubenswrapper[4987]: I1004 07:27:14.048058 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" event={"ID":"70c28efa-16a5-4eef-8fec-f80bad6cea42","Type":"ContainerDied","Data":"d8d0eed68c5cb89f47a62782bbc3fc4a689f678386e9193651221899eccd6ee5"} Oct 04 07:27:15 crc kubenswrapper[4987]: I1004 07:27:15.058004 4987 generic.go:334] "Generic (PLEG): container finished" podID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerID="dc255fa7fd733647124e6a8c5a98d8e24c6daa6f7d600b210a892335357dc0ea" exitCode=0 Oct 04 07:27:15 crc kubenswrapper[4987]: I1004 07:27:15.058128 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" event={"ID":"70c28efa-16a5-4eef-8fec-f80bad6cea42","Type":"ContainerDied","Data":"dc255fa7fd733647124e6a8c5a98d8e24c6daa6f7d600b210a892335357dc0ea"} Oct 04 07:27:15 crc kubenswrapper[4987]: I1004 07:27:15.061342 4987 generic.go:334] "Generic (PLEG): container finished" podID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerID="5cf13815cd5e3b982434f0cfed358c72f5d00802e83fff45712331a55f736988" exitCode=0 Oct 04 07:27:15 crc kubenswrapper[4987]: I1004 07:27:15.061434 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" event={"ID":"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3","Type":"ContainerDied","Data":"5cf13815cd5e3b982434f0cfed358c72f5d00802e83fff45712331a55f736988"} Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.069293 4987 generic.go:334] "Generic (PLEG): container finished" podID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerID="b569c0a5223fbef78608ed50ee55a6ab57f5ff78e29bf8a1df1eb86a607e6809" exitCode=0 Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.069385 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" event={"ID":"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3","Type":"ContainerDied","Data":"b569c0a5223fbef78608ed50ee55a6ab57f5ff78e29bf8a1df1eb86a607e6809"} Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.070658 4987 generic.go:334] "Generic (PLEG): container finished" podID="91962d41-2412-432a-823e-a4351de82e52" containerID="f50f5cdd0a355c36be1e6022a1de82292de621a28d536a0f7bb0735e03d1e5be" exitCode=0 Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.070837 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" event={"ID":"91962d41-2412-432a-823e-a4351de82e52","Type":"ContainerDied","Data":"f50f5cdd0a355c36be1e6022a1de82292de621a28d536a0f7bb0735e03d1e5be"} Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.320385 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.446008 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-bundle\") pod \"70c28efa-16a5-4eef-8fec-f80bad6cea42\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.446051 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-util\") pod \"70c28efa-16a5-4eef-8fec-f80bad6cea42\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.446174 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhd8v\" (UniqueName: \"kubernetes.io/projected/70c28efa-16a5-4eef-8fec-f80bad6cea42-kube-api-access-mhd8v\") pod \"70c28efa-16a5-4eef-8fec-f80bad6cea42\" (UID: \"70c28efa-16a5-4eef-8fec-f80bad6cea42\") " Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.448344 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-bundle" (OuterVolumeSpecName: "bundle") pod "70c28efa-16a5-4eef-8fec-f80bad6cea42" (UID: "70c28efa-16a5-4eef-8fec-f80bad6cea42"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.457804 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c28efa-16a5-4eef-8fec-f80bad6cea42-kube-api-access-mhd8v" (OuterVolumeSpecName: "kube-api-access-mhd8v") pod "70c28efa-16a5-4eef-8fec-f80bad6cea42" (UID: "70c28efa-16a5-4eef-8fec-f80bad6cea42"). InnerVolumeSpecName "kube-api-access-mhd8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.487859 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-util" (OuterVolumeSpecName: "util") pod "70c28efa-16a5-4eef-8fec-f80bad6cea42" (UID: "70c28efa-16a5-4eef-8fec-f80bad6cea42"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.547400 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.547442 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70c28efa-16a5-4eef-8fec-f80bad6cea42-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:16 crc kubenswrapper[4987]: I1004 07:27:16.547451 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhd8v\" (UniqueName: \"kubernetes.io/projected/70c28efa-16a5-4eef-8fec-f80bad6cea42-kube-api-access-mhd8v\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.080779 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" event={"ID":"70c28efa-16a5-4eef-8fec-f80bad6cea42","Type":"ContainerDied","Data":"4971f943a12341787c681feb1795ecf731da9265738bf5fc36e52d4439a7a7fd"} Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.081761 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4971f943a12341787c681feb1795ecf731da9265738bf5fc36e52d4439a7a7fd" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.080929 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.436617 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.457350 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.482893 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjn5m\" (UniqueName: \"kubernetes.io/projected/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-kube-api-access-qjn5m\") pod \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.483007 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91962d41-2412-432a-823e-a4351de82e52-config-data\") pod \"91962d41-2412-432a-823e-a4351de82e52\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.483108 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2gxf\" (UniqueName: \"kubernetes.io/projected/91962d41-2412-432a-823e-a4351de82e52-kube-api-access-k2gxf\") pod \"91962d41-2412-432a-823e-a4351de82e52\" (UID: \"91962d41-2412-432a-823e-a4351de82e52\") " Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.483165 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-util\") pod \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.483233 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-bundle\") pod \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\" (UID: \"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3\") " Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.485463 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-bundle" (OuterVolumeSpecName: "bundle") pod "a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" (UID: "a5d0cde7-7997-4601-b1e1-f8d55d28f8b3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.491550 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-kube-api-access-qjn5m" (OuterVolumeSpecName: "kube-api-access-qjn5m") pod "a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" (UID: "a5d0cde7-7997-4601-b1e1-f8d55d28f8b3"). InnerVolumeSpecName "kube-api-access-qjn5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.492331 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91962d41-2412-432a-823e-a4351de82e52-kube-api-access-k2gxf" (OuterVolumeSpecName: "kube-api-access-k2gxf") pod "91962d41-2412-432a-823e-a4351de82e52" (UID: "91962d41-2412-432a-823e-a4351de82e52"). InnerVolumeSpecName "kube-api-access-k2gxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.499078 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-util" (OuterVolumeSpecName: "util") pod "a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" (UID: "a5d0cde7-7997-4601-b1e1-f8d55d28f8b3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.528832 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91962d41-2412-432a-823e-a4351de82e52-config-data" (OuterVolumeSpecName: "config-data") pod "91962d41-2412-432a-823e-a4351de82e52" (UID: "91962d41-2412-432a-823e-a4351de82e52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.586390 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjn5m\" (UniqueName: \"kubernetes.io/projected/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-kube-api-access-qjn5m\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.586449 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91962d41-2412-432a-823e-a4351de82e52-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.586467 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2gxf\" (UniqueName: \"kubernetes.io/projected/91962d41-2412-432a-823e-a4351de82e52-kube-api-access-k2gxf\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.586479 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:17 crc kubenswrapper[4987]: I1004 07:27:17.586494 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5d0cde7-7997-4601-b1e1-f8d55d28f8b3-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.092925 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" event={"ID":"a5d0cde7-7997-4601-b1e1-f8d55d28f8b3","Type":"ContainerDied","Data":"7492e7fdc77249d0e291fdcf638961f03e9a3d74f6a5964ee307f6c73b14743f"} Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.092977 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7492e7fdc77249d0e291fdcf638961f03e9a3d74f6a5964ee307f6c73b14743f" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.093026 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.097451 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" event={"ID":"91962d41-2412-432a-823e-a4351de82e52","Type":"ContainerDied","Data":"c67b22cb5e8ca3a2293e8eaa88fe7090101b06ef92cf835c99a435e831f76fdc"} Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.097695 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c67b22cb5e8ca3a2293e8eaa88fe7090101b06ef92cf835c99a435e831f76fdc" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.097529 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-6sb9w" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.314550 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-c27lb"] Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.314934 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="extract" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.314958 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="extract" Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.314977 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="extract" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.314987 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="extract" Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.314998 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="util" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315007 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="util" Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.315023 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="pull" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315031 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="pull" Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.315045 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="util" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315054 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="util" Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.315071 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91962d41-2412-432a-823e-a4351de82e52" containerName="keystone-db-sync" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315081 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="91962d41-2412-432a-823e-a4351de82e52" containerName="keystone-db-sync" Oct 04 07:27:18 crc kubenswrapper[4987]: E1004 07:27:18.315096 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="pull" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315104 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="pull" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315251 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c28efa-16a5-4eef-8fec-f80bad6cea42" containerName="extract" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315263 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d0cde7-7997-4601-b1e1-f8d55d28f8b3" containerName="extract" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315284 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="91962d41-2412-432a-823e-a4351de82e52" containerName="keystone-db-sync" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.315920 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.319047 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.319053 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-l82gw" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.319421 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.319509 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.331668 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-c27lb"] Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.401115 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-config-data\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.401178 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkmg9\" (UniqueName: \"kubernetes.io/projected/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-kube-api-access-xkmg9\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.401211 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-scripts\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.401337 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-credential-keys\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.401544 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-fernet-keys\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.503515 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-fernet-keys\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.504336 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-config-data\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.504433 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkmg9\" (UniqueName: \"kubernetes.io/projected/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-kube-api-access-xkmg9\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.504530 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-scripts\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.504641 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-credential-keys\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.513792 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-fernet-keys\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.514558 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-config-data\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.514600 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-scripts\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.516316 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-credential-keys\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.525283 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkmg9\" (UniqueName: \"kubernetes.io/projected/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-kube-api-access-xkmg9\") pod \"keystone-bootstrap-c27lb\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.632785 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:18 crc kubenswrapper[4987]: I1004 07:27:18.847563 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-c27lb"] Oct 04 07:27:19 crc kubenswrapper[4987]: I1004 07:27:19.107037 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" event={"ID":"814ffbb2-3f6a-4658-9e65-9c592d4d5b92","Type":"ContainerStarted","Data":"d90ead49c8426aa3aee98ecb42a5e8a0b92d1c9f822ec4e481ab244c9ee1aead"} Oct 04 07:27:19 crc kubenswrapper[4987]: I1004 07:27:19.107387 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" event={"ID":"814ffbb2-3f6a-4658-9e65-9c592d4d5b92","Type":"ContainerStarted","Data":"3b7b2310b6366e8ceb6caa8f2ef1eaa2312d5b0095a03623216c6ece346844e7"} Oct 04 07:27:19 crc kubenswrapper[4987]: I1004 07:27:19.126373 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" podStartSLOduration=1.126352518 podStartE2EDuration="1.126352518s" podCreationTimestamp="2025-10-04 07:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:27:19.121388893 +0000 UTC m=+1056.176287735" watchObservedRunningTime="2025-10-04 07:27:19.126352518 +0000 UTC m=+1056.181251360" Oct 04 07:27:24 crc kubenswrapper[4987]: I1004 07:27:24.160412 4987 generic.go:334] "Generic (PLEG): container finished" podID="814ffbb2-3f6a-4658-9e65-9c592d4d5b92" containerID="d90ead49c8426aa3aee98ecb42a5e8a0b92d1c9f822ec4e481ab244c9ee1aead" exitCode=0 Oct 04 07:27:24 crc kubenswrapper[4987]: I1004 07:27:24.160517 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" event={"ID":"814ffbb2-3f6a-4658-9e65-9c592d4d5b92","Type":"ContainerDied","Data":"d90ead49c8426aa3aee98ecb42a5e8a0b92d1c9f822ec4e481ab244c9ee1aead"} Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.472939 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.504916 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-scripts\") pod \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.505022 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-config-data\") pod \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.505911 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-fernet-keys\") pod \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.506861 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-credential-keys\") pod \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.506921 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkmg9\" (UniqueName: \"kubernetes.io/projected/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-kube-api-access-xkmg9\") pod \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\" (UID: \"814ffbb2-3f6a-4658-9e65-9c592d4d5b92\") " Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.512087 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-kube-api-access-xkmg9" (OuterVolumeSpecName: "kube-api-access-xkmg9") pod "814ffbb2-3f6a-4658-9e65-9c592d4d5b92" (UID: "814ffbb2-3f6a-4658-9e65-9c592d4d5b92"). InnerVolumeSpecName "kube-api-access-xkmg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.512165 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-scripts" (OuterVolumeSpecName: "scripts") pod "814ffbb2-3f6a-4658-9e65-9c592d4d5b92" (UID: "814ffbb2-3f6a-4658-9e65-9c592d4d5b92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.514719 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "814ffbb2-3f6a-4658-9e65-9c592d4d5b92" (UID: "814ffbb2-3f6a-4658-9e65-9c592d4d5b92"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.520221 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "814ffbb2-3f6a-4658-9e65-9c592d4d5b92" (UID: "814ffbb2-3f6a-4658-9e65-9c592d4d5b92"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.530919 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-config-data" (OuterVolumeSpecName: "config-data") pod "814ffbb2-3f6a-4658-9e65-9c592d4d5b92" (UID: "814ffbb2-3f6a-4658-9e65-9c592d4d5b92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.609653 4987 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.609692 4987 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.609704 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkmg9\" (UniqueName: \"kubernetes.io/projected/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-kube-api-access-xkmg9\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.609716 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:25 crc kubenswrapper[4987]: I1004 07:27:25.609725 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814ffbb2-3f6a-4658-9e65-9c592d4d5b92-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.174575 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" event={"ID":"814ffbb2-3f6a-4658-9e65-9c592d4d5b92","Type":"ContainerDied","Data":"3b7b2310b6366e8ceb6caa8f2ef1eaa2312d5b0095a03623216c6ece346844e7"} Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.174933 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b7b2310b6366e8ceb6caa8f2ef1eaa2312d5b0095a03623216c6ece346844e7" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.174684 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-c27lb" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.250483 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-6fccb96676-gqppm"] Oct 04 07:27:26 crc kubenswrapper[4987]: E1004 07:27:26.250998 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="814ffbb2-3f6a-4658-9e65-9c592d4d5b92" containerName="keystone-bootstrap" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.251032 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="814ffbb2-3f6a-4658-9e65-9c592d4d5b92" containerName="keystone-bootstrap" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.251298 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="814ffbb2-3f6a-4658-9e65-9c592d4d5b92" containerName="keystone-bootstrap" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.252418 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.255742 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.255749 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-l82gw" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.256009 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.257469 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.266966 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-6fccb96676-gqppm"] Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.320827 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-fernet-keys\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.321017 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw5zg\" (UniqueName: \"kubernetes.io/projected/4851c21f-3f97-4813-b27f-64d7e7e2b764-kube-api-access-cw5zg\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.321107 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-credential-keys\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.321144 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-config-data\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.321188 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-scripts\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.422943 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw5zg\" (UniqueName: \"kubernetes.io/projected/4851c21f-3f97-4813-b27f-64d7e7e2b764-kube-api-access-cw5zg\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.423002 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-credential-keys\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.423042 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-config-data\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.423067 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-scripts\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.423102 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-fernet-keys\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.427913 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-fernet-keys\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.428440 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-config-data\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.428644 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-credential-keys\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.433189 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4851c21f-3f97-4813-b27f-64d7e7e2b764-scripts\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.446493 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw5zg\" (UniqueName: \"kubernetes.io/projected/4851c21f-3f97-4813-b27f-64d7e7e2b764-kube-api-access-cw5zg\") pod \"keystone-6fccb96676-gqppm\" (UID: \"4851c21f-3f97-4813-b27f-64d7e7e2b764\") " pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:26 crc kubenswrapper[4987]: I1004 07:27:26.594424 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:27 crc kubenswrapper[4987]: I1004 07:27:27.167264 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-6fccb96676-gqppm"] Oct 04 07:27:27 crc kubenswrapper[4987]: I1004 07:27:27.185654 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" event={"ID":"4851c21f-3f97-4813-b27f-64d7e7e2b764","Type":"ContainerStarted","Data":"40394d935e98da415879f25bf5a72c8b3c03df3bdd1aad456b516d0b4ddb007a"} Oct 04 07:27:28 crc kubenswrapper[4987]: I1004 07:27:28.194585 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" event={"ID":"4851c21f-3f97-4813-b27f-64d7e7e2b764","Type":"ContainerStarted","Data":"fedeccdb29df53d8c86fe2e477b534de7b54e6e774577685bf9dae62646abf34"} Oct 04 07:27:28 crc kubenswrapper[4987]: I1004 07:27:28.195361 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:27:28 crc kubenswrapper[4987]: I1004 07:27:28.219807 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" podStartSLOduration=2.219784676 podStartE2EDuration="2.219784676s" podCreationTimestamp="2025-10-04 07:27:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:27:28.216245526 +0000 UTC m=+1065.271144378" watchObservedRunningTime="2025-10-04 07:27:28.219784676 +0000 UTC m=+1065.274683518" Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.834295 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x"] Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.836356 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.839937 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.840512 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-z8gdq" Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.894232 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x"] Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.947261 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l7vr\" (UniqueName: \"kubernetes.io/projected/a0639060-01de-41cc-90b7-e1cde85170fc-kube-api-access-8l7vr\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.947319 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0639060-01de-41cc-90b7-e1cde85170fc-apiservice-cert\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:33 crc kubenswrapper[4987]: I1004 07:27:33.947351 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0639060-01de-41cc-90b7-e1cde85170fc-webhook-cert\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.048652 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l7vr\" (UniqueName: \"kubernetes.io/projected/a0639060-01de-41cc-90b7-e1cde85170fc-kube-api-access-8l7vr\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.048930 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0639060-01de-41cc-90b7-e1cde85170fc-apiservice-cert\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.048970 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0639060-01de-41cc-90b7-e1cde85170fc-webhook-cert\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.056683 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a0639060-01de-41cc-90b7-e1cde85170fc-webhook-cert\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.068175 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a0639060-01de-41cc-90b7-e1cde85170fc-apiservice-cert\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.070202 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l7vr\" (UniqueName: \"kubernetes.io/projected/a0639060-01de-41cc-90b7-e1cde85170fc-kube-api-access-8l7vr\") pod \"horizon-operator-controller-manager-687c78db9b-kfz5x\" (UID: \"a0639060-01de-41cc-90b7-e1cde85170fc\") " pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.159345 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:34 crc kubenswrapper[4987]: I1004 07:27:34.630899 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x"] Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.259497 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" event={"ID":"a0639060-01de-41cc-90b7-e1cde85170fc","Type":"ContainerStarted","Data":"7956ffd3f0e5f9b6343f156add56fa2f7987d9ac635b4a7189e44a0f483ed505"} Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.639614 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc"] Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.642047 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.644392 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.644397 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wj28h" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.652433 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc"] Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.777315 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d3c896ee-c16b-47c4-b74a-706168509fb5-apiservice-cert\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.777417 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d3c896ee-c16b-47c4-b74a-706168509fb5-webhook-cert\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.777461 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqwmz\" (UniqueName: \"kubernetes.io/projected/d3c896ee-c16b-47c4-b74a-706168509fb5-kube-api-access-fqwmz\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.879248 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d3c896ee-c16b-47c4-b74a-706168509fb5-apiservice-cert\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.879389 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d3c896ee-c16b-47c4-b74a-706168509fb5-webhook-cert\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.879434 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqwmz\" (UniqueName: \"kubernetes.io/projected/d3c896ee-c16b-47c4-b74a-706168509fb5-kube-api-access-fqwmz\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.888007 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d3c896ee-c16b-47c4-b74a-706168509fb5-apiservice-cert\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.889058 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d3c896ee-c16b-47c4-b74a-706168509fb5-webhook-cert\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.917533 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqwmz\" (UniqueName: \"kubernetes.io/projected/d3c896ee-c16b-47c4-b74a-706168509fb5-kube-api-access-fqwmz\") pod \"swift-operator-controller-manager-6dcf9c964f-9rbnc\" (UID: \"d3c896ee-c16b-47c4-b74a-706168509fb5\") " pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:35 crc kubenswrapper[4987]: I1004 07:27:35.965912 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:36 crc kubenswrapper[4987]: I1004 07:27:36.478007 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc"] Oct 04 07:27:36 crc kubenswrapper[4987]: W1004 07:27:36.487228 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3c896ee_c16b_47c4_b74a_706168509fb5.slice/crio-f02dcd2758b94a500780305af376aeb6b0161786a03de6d8a9b9e2e0073aaed1 WatchSource:0}: Error finding container f02dcd2758b94a500780305af376aeb6b0161786a03de6d8a9b9e2e0073aaed1: Status 404 returned error can't find the container with id f02dcd2758b94a500780305af376aeb6b0161786a03de6d8a9b9e2e0073aaed1 Oct 04 07:27:37 crc kubenswrapper[4987]: I1004 07:27:37.285214 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" event={"ID":"d3c896ee-c16b-47c4-b74a-706168509fb5","Type":"ContainerStarted","Data":"f02dcd2758b94a500780305af376aeb6b0161786a03de6d8a9b9e2e0073aaed1"} Oct 04 07:27:41 crc kubenswrapper[4987]: I1004 07:27:41.340729 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" event={"ID":"d3c896ee-c16b-47c4-b74a-706168509fb5","Type":"ContainerStarted","Data":"50a85c38da87324a4340266d98fa708c08f310fea3209a39b7b94e985bed3e87"} Oct 04 07:27:41 crc kubenswrapper[4987]: I1004 07:27:41.341528 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:41 crc kubenswrapper[4987]: I1004 07:27:41.341541 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" event={"ID":"d3c896ee-c16b-47c4-b74a-706168509fb5","Type":"ContainerStarted","Data":"1377917b71d30673acd7ca7cc2cfffe91ad024903048811d022d61b70f1eeccd"} Oct 04 07:27:42 crc kubenswrapper[4987]: I1004 07:27:42.350683 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" event={"ID":"a0639060-01de-41cc-90b7-e1cde85170fc","Type":"ContainerStarted","Data":"2e498041f5c796a96f3df70ae8755beac7cbedba777f1b914e33e51a065be208"} Oct 04 07:27:42 crc kubenswrapper[4987]: I1004 07:27:42.351269 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" event={"ID":"a0639060-01de-41cc-90b7-e1cde85170fc","Type":"ContainerStarted","Data":"9897bf945818b87b22ee97dc1614085169a90fa97bd45a8788f162de7b97c1f3"} Oct 04 07:27:42 crc kubenswrapper[4987]: I1004 07:27:42.378268 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" podStartSLOduration=2.401933562 podStartE2EDuration="9.378245611s" podCreationTimestamp="2025-10-04 07:27:33 +0000 UTC" firstStartedPulling="2025-10-04 07:27:34.64060421 +0000 UTC m=+1071.695503052" lastFinishedPulling="2025-10-04 07:27:41.616916259 +0000 UTC m=+1078.671815101" observedRunningTime="2025-10-04 07:27:42.376507787 +0000 UTC m=+1079.431406629" watchObservedRunningTime="2025-10-04 07:27:42.378245611 +0000 UTC m=+1079.433144453" Oct 04 07:27:42 crc kubenswrapper[4987]: I1004 07:27:42.379109 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" podStartSLOduration=3.306188637 podStartE2EDuration="7.379098717s" podCreationTimestamp="2025-10-04 07:27:35 +0000 UTC" firstStartedPulling="2025-10-04 07:27:36.491878504 +0000 UTC m=+1073.546777346" lastFinishedPulling="2025-10-04 07:27:40.564788584 +0000 UTC m=+1077.619687426" observedRunningTime="2025-10-04 07:27:41.368905998 +0000 UTC m=+1078.423804860" watchObservedRunningTime="2025-10-04 07:27:42.379098717 +0000 UTC m=+1079.433997579" Oct 04 07:27:43 crc kubenswrapper[4987]: I1004 07:27:43.359732 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:45 crc kubenswrapper[4987]: I1004 07:27:45.971861 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6dcf9c964f-9rbnc" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.710268 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.715689 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.719578 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.719885 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.720371 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-7pkcd" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.724187 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.754680 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.813268 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgcdn\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-kube-api-access-vgcdn\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.813352 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/70e18ecc-8da8-4423-8eff-39f07bf2383f-lock\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.813385 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/70e18ecc-8da8-4423-8eff-39f07bf2383f-cache\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.813504 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.813569 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.915493 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgcdn\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-kube-api-access-vgcdn\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.915581 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/70e18ecc-8da8-4423-8eff-39f07bf2383f-lock\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.915607 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/70e18ecc-8da8-4423-8eff-39f07bf2383f-cache\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.915684 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.915734 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: E1004 07:27:48.915938 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:27:48 crc kubenswrapper[4987]: E1004 07:27:48.915962 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 07:27:48 crc kubenswrapper[4987]: E1004 07:27:48.916040 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift podName:70e18ecc-8da8-4423-8eff-39f07bf2383f nodeName:}" failed. No retries permitted until 2025-10-04 07:27:49.416010405 +0000 UTC m=+1086.470909247 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift") pod "swift-storage-0" (UID: "70e18ecc-8da8-4423-8eff-39f07bf2383f") : configmap "swift-ring-files" not found Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.916105 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.916369 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/70e18ecc-8da8-4423-8eff-39f07bf2383f-cache\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.916664 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/70e18ecc-8da8-4423-8eff-39f07bf2383f-lock\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.940182 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:48 crc kubenswrapper[4987]: I1004 07:27:48.942033 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgcdn\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-kube-api-access-vgcdn\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:49 crc kubenswrapper[4987]: I1004 07:27:49.425261 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:49 crc kubenswrapper[4987]: E1004 07:27:49.425662 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:27:49 crc kubenswrapper[4987]: E1004 07:27:49.425724 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 07:27:49 crc kubenswrapper[4987]: E1004 07:27:49.425858 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift podName:70e18ecc-8da8-4423-8eff-39f07bf2383f nodeName:}" failed. No retries permitted until 2025-10-04 07:27:50.425812696 +0000 UTC m=+1087.480711568 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift") pod "swift-storage-0" (UID: "70e18ecc-8da8-4423-8eff-39f07bf2383f") : configmap "swift-ring-files" not found Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.303388 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-rzcdc"] Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.304526 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.307065 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-9ff2n" Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.315717 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-rzcdc"] Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.443662 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4snsz\" (UniqueName: \"kubernetes.io/projected/df486065-5174-4d53-a6aa-ba2a80fd7354-kube-api-access-4snsz\") pod \"glance-operator-index-rzcdc\" (UID: \"df486065-5174-4d53-a6aa-ba2a80fd7354\") " pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.444224 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:50 crc kubenswrapper[4987]: E1004 07:27:50.444506 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:27:50 crc kubenswrapper[4987]: E1004 07:27:50.444560 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 07:27:50 crc kubenswrapper[4987]: E1004 07:27:50.444653 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift podName:70e18ecc-8da8-4423-8eff-39f07bf2383f nodeName:}" failed. No retries permitted until 2025-10-04 07:27:52.444612274 +0000 UTC m=+1089.499511116 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift") pod "swift-storage-0" (UID: "70e18ecc-8da8-4423-8eff-39f07bf2383f") : configmap "swift-ring-files" not found Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.545910 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4snsz\" (UniqueName: \"kubernetes.io/projected/df486065-5174-4d53-a6aa-ba2a80fd7354-kube-api-access-4snsz\") pod \"glance-operator-index-rzcdc\" (UID: \"df486065-5174-4d53-a6aa-ba2a80fd7354\") " pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.566745 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4snsz\" (UniqueName: \"kubernetes.io/projected/df486065-5174-4d53-a6aa-ba2a80fd7354-kube-api-access-4snsz\") pod \"glance-operator-index-rzcdc\" (UID: \"df486065-5174-4d53-a6aa-ba2a80fd7354\") " pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.628917 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:27:50 crc kubenswrapper[4987]: I1004 07:27:50.976170 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-rzcdc"] Oct 04 07:27:51 crc kubenswrapper[4987]: I1004 07:27:51.438427 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-rzcdc" event={"ID":"df486065-5174-4d53-a6aa-ba2a80fd7354","Type":"ContainerStarted","Data":"9063dbeb545e12996a8e246700968eb079192b518703668125bf92054046ce97"} Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.478766 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:52 crc kubenswrapper[4987]: E1004 07:27:52.479003 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:27:52 crc kubenswrapper[4987]: E1004 07:27:52.479022 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 07:27:52 crc kubenswrapper[4987]: E1004 07:27:52.479076 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift podName:70e18ecc-8da8-4423-8eff-39f07bf2383f nodeName:}" failed. No retries permitted until 2025-10-04 07:27:56.479058221 +0000 UTC m=+1093.533957063 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift") pod "swift-storage-0" (UID: "70e18ecc-8da8-4423-8eff-39f07bf2383f") : configmap "swift-ring-files" not found Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.773486 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-zj4nq"] Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.774457 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.780230 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.780549 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.780965 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.793798 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-zj4nq"] Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.887439 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a531f5c9-a54b-4016-83d4-57e4ced832d6-etc-swift\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.887494 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z28wq\" (UniqueName: \"kubernetes.io/projected/a531f5c9-a54b-4016-83d4-57e4ced832d6-kube-api-access-z28wq\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.887525 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-dispersionconf\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.887559 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-scripts\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.887578 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-ring-data-devices\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.887600 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-swiftconf\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.989397 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a531f5c9-a54b-4016-83d4-57e4ced832d6-etc-swift\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.989447 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z28wq\" (UniqueName: \"kubernetes.io/projected/a531f5c9-a54b-4016-83d4-57e4ced832d6-kube-api-access-z28wq\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.989538 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-dispersionconf\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.989569 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-scripts\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.989588 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-ring-data-devices\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.989616 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-swiftconf\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.990590 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-scripts\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.990619 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-ring-data-devices\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.990615 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a531f5c9-a54b-4016-83d4-57e4ced832d6-etc-swift\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.996427 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-swiftconf\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:52 crc kubenswrapper[4987]: I1004 07:27:52.996844 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-dispersionconf\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:53 crc kubenswrapper[4987]: I1004 07:27:53.012575 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z28wq\" (UniqueName: \"kubernetes.io/projected/a531f5c9-a54b-4016-83d4-57e4ced832d6-kube-api-access-z28wq\") pod \"swift-ring-rebalance-zj4nq\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:53 crc kubenswrapper[4987]: I1004 07:27:53.095233 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:27:53 crc kubenswrapper[4987]: I1004 07:27:53.622407 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-zj4nq"] Oct 04 07:27:53 crc kubenswrapper[4987]: W1004 07:27:53.662100 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda531f5c9_a54b_4016_83d4_57e4ced832d6.slice/crio-086653578db9860abb8f523ca03f70b8c3e7b0028e6ee045deb7a071a1981522 WatchSource:0}: Error finding container 086653578db9860abb8f523ca03f70b8c3e7b0028e6ee045deb7a071a1981522: Status 404 returned error can't find the container with id 086653578db9860abb8f523ca03f70b8c3e7b0028e6ee045deb7a071a1981522 Oct 04 07:27:54 crc kubenswrapper[4987]: I1004 07:27:54.166043 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-687c78db9b-kfz5x" Oct 04 07:27:54 crc kubenswrapper[4987]: I1004 07:27:54.465983 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" event={"ID":"a531f5c9-a54b-4016-83d4-57e4ced832d6","Type":"ContainerStarted","Data":"086653578db9860abb8f523ca03f70b8c3e7b0028e6ee045deb7a071a1981522"} Oct 04 07:27:55 crc kubenswrapper[4987]: I1004 07:27:55.495451 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-rzcdc"] Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.098005 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-f5msg"] Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.100079 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.122729 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-f5msg"] Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.259426 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfkjh\" (UniqueName: \"kubernetes.io/projected/21c12a08-bede-4878-b1ad-3f847091bf9a-kube-api-access-zfkjh\") pod \"glance-operator-index-f5msg\" (UID: \"21c12a08-bede-4878-b1ad-3f847091bf9a\") " pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.361732 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfkjh\" (UniqueName: \"kubernetes.io/projected/21c12a08-bede-4878-b1ad-3f847091bf9a-kube-api-access-zfkjh\") pod \"glance-operator-index-f5msg\" (UID: \"21c12a08-bede-4878-b1ad-3f847091bf9a\") " pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.382328 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfkjh\" (UniqueName: \"kubernetes.io/projected/21c12a08-bede-4878-b1ad-3f847091bf9a-kube-api-access-zfkjh\") pod \"glance-operator-index-f5msg\" (UID: \"21c12a08-bede-4878-b1ad-3f847091bf9a\") " pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.419723 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:27:56 crc kubenswrapper[4987]: I1004 07:27:56.565675 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:27:56 crc kubenswrapper[4987]: E1004 07:27:56.566061 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:27:56 crc kubenswrapper[4987]: E1004 07:27:56.566101 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 07:27:56 crc kubenswrapper[4987]: E1004 07:27:56.566177 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift podName:70e18ecc-8da8-4423-8eff-39f07bf2383f nodeName:}" failed. No retries permitted until 2025-10-04 07:28:04.566157181 +0000 UTC m=+1101.621056023 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift") pod "swift-storage-0" (UID: "70e18ecc-8da8-4423-8eff-39f07bf2383f") : configmap "swift-ring-files" not found Oct 04 07:27:58 crc kubenswrapper[4987]: I1004 07:27:58.091815 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-6fccb96676-gqppm" Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.531362 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-rzcdc" event={"ID":"df486065-5174-4d53-a6aa-ba2a80fd7354","Type":"ContainerStarted","Data":"3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b"} Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.531729 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-rzcdc" podUID="df486065-5174-4d53-a6aa-ba2a80fd7354" containerName="registry-server" containerID="cri-o://3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b" gracePeriod=2 Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.534682 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" event={"ID":"a531f5c9-a54b-4016-83d4-57e4ced832d6","Type":"ContainerStarted","Data":"345d5d44dbc11c3bafc6bb41166f7fe2b99c97df85013c5d3cc109bf4216450a"} Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.558372 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-rzcdc" podStartSLOduration=1.378073516 podStartE2EDuration="11.558347039s" podCreationTimestamp="2025-10-04 07:27:50 +0000 UTC" firstStartedPulling="2025-10-04 07:27:50.98891069 +0000 UTC m=+1088.043809532" lastFinishedPulling="2025-10-04 07:28:01.169184213 +0000 UTC m=+1098.224083055" observedRunningTime="2025-10-04 07:28:01.550978409 +0000 UTC m=+1098.605877251" watchObservedRunningTime="2025-10-04 07:28:01.558347039 +0000 UTC m=+1098.613245891" Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.569387 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" podStartSLOduration=1.9029820499999999 podStartE2EDuration="9.569366681s" podCreationTimestamp="2025-10-04 07:27:52 +0000 UTC" firstStartedPulling="2025-10-04 07:27:53.667740557 +0000 UTC m=+1090.722639399" lastFinishedPulling="2025-10-04 07:28:01.334125188 +0000 UTC m=+1098.389024030" observedRunningTime="2025-10-04 07:28:01.567881726 +0000 UTC m=+1098.622780568" watchObservedRunningTime="2025-10-04 07:28:01.569366681 +0000 UTC m=+1098.624265533" Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.681025 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-f5msg"] Oct 04 07:28:01 crc kubenswrapper[4987]: W1004 07:28:01.689839 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21c12a08_bede_4878_b1ad_3f847091bf9a.slice/crio-ca57013bdd8e8f08d742c8721b8703ca7b91e331544401e440ae7fe6ee0a41ff WatchSource:0}: Error finding container ca57013bdd8e8f08d742c8721b8703ca7b91e331544401e440ae7fe6ee0a41ff: Status 404 returned error can't find the container with id ca57013bdd8e8f08d742c8721b8703ca7b91e331544401e440ae7fe6ee0a41ff Oct 04 07:28:01 crc kubenswrapper[4987]: I1004 07:28:01.904306 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.063221 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4snsz\" (UniqueName: \"kubernetes.io/projected/df486065-5174-4d53-a6aa-ba2a80fd7354-kube-api-access-4snsz\") pod \"df486065-5174-4d53-a6aa-ba2a80fd7354\" (UID: \"df486065-5174-4d53-a6aa-ba2a80fd7354\") " Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.067589 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df486065-5174-4d53-a6aa-ba2a80fd7354-kube-api-access-4snsz" (OuterVolumeSpecName: "kube-api-access-4snsz") pod "df486065-5174-4d53-a6aa-ba2a80fd7354" (UID: "df486065-5174-4d53-a6aa-ba2a80fd7354"). InnerVolumeSpecName "kube-api-access-4snsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.165331 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4snsz\" (UniqueName: \"kubernetes.io/projected/df486065-5174-4d53-a6aa-ba2a80fd7354-kube-api-access-4snsz\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.542316 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-f5msg" event={"ID":"21c12a08-bede-4878-b1ad-3f847091bf9a","Type":"ContainerStarted","Data":"bd3b48483e2cd398ac3204cd82de94502b8b937a62ff857b82fdee04592591d8"} Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.542642 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-f5msg" event={"ID":"21c12a08-bede-4878-b1ad-3f847091bf9a","Type":"ContainerStarted","Data":"ca57013bdd8e8f08d742c8721b8703ca7b91e331544401e440ae7fe6ee0a41ff"} Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.544076 4987 generic.go:334] "Generic (PLEG): container finished" podID="df486065-5174-4d53-a6aa-ba2a80fd7354" containerID="3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b" exitCode=0 Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.544135 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-rzcdc" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.544132 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-rzcdc" event={"ID":"df486065-5174-4d53-a6aa-ba2a80fd7354","Type":"ContainerDied","Data":"3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b"} Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.544194 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-rzcdc" event={"ID":"df486065-5174-4d53-a6aa-ba2a80fd7354","Type":"ContainerDied","Data":"9063dbeb545e12996a8e246700968eb079192b518703668125bf92054046ce97"} Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.544218 4987 scope.go:117] "RemoveContainer" containerID="3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.559735 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-f5msg" podStartSLOduration=6.496806555 podStartE2EDuration="6.559712113s" podCreationTimestamp="2025-10-04 07:27:56 +0000 UTC" firstStartedPulling="2025-10-04 07:28:01.694034403 +0000 UTC m=+1098.748933245" lastFinishedPulling="2025-10-04 07:28:01.756939951 +0000 UTC m=+1098.811838803" observedRunningTime="2025-10-04 07:28:02.557751742 +0000 UTC m=+1099.612650584" watchObservedRunningTime="2025-10-04 07:28:02.559712113 +0000 UTC m=+1099.614610955" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.569282 4987 scope.go:117] "RemoveContainer" containerID="3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b" Oct 04 07:28:02 crc kubenswrapper[4987]: E1004 07:28:02.570729 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b\": container with ID starting with 3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b not found: ID does not exist" containerID="3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.570781 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b"} err="failed to get container status \"3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b\": rpc error: code = NotFound desc = could not find container \"3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b\": container with ID starting with 3b35d365bd75dac9d78a940d0c5076cb3bd85f310fe1ceb2852dbe51a277152b not found: ID does not exist" Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.584277 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-rzcdc"] Oct 04 07:28:02 crc kubenswrapper[4987]: I1004 07:28:02.587879 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-rzcdc"] Oct 04 07:28:04 crc kubenswrapper[4987]: I1004 07:28:04.064134 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df486065-5174-4d53-a6aa-ba2a80fd7354" path="/var/lib/kubelet/pods/df486065-5174-4d53-a6aa-ba2a80fd7354/volumes" Oct 04 07:28:04 crc kubenswrapper[4987]: I1004 07:28:04.606260 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:28:04 crc kubenswrapper[4987]: E1004 07:28:04.606480 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:28:04 crc kubenswrapper[4987]: E1004 07:28:04.606500 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Oct 04 07:28:04 crc kubenswrapper[4987]: E1004 07:28:04.606558 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift podName:70e18ecc-8da8-4423-8eff-39f07bf2383f nodeName:}" failed. No retries permitted until 2025-10-04 07:28:20.606535615 +0000 UTC m=+1117.661434457 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift") pod "swift-storage-0" (UID: "70e18ecc-8da8-4423-8eff-39f07bf2383f") : configmap "swift-ring-files" not found Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.460991 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9"] Oct 04 07:28:05 crc kubenswrapper[4987]: E1004 07:28:05.461320 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df486065-5174-4d53-a6aa-ba2a80fd7354" containerName="registry-server" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.461333 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="df486065-5174-4d53-a6aa-ba2a80fd7354" containerName="registry-server" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.461484 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="df486065-5174-4d53-a6aa-ba2a80fd7354" containerName="registry-server" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.462230 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.477798 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9"] Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.622677 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkdx2\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-kube-api-access-kkdx2\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.622836 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.622993 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-config-data\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.623086 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-run-httpd\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.623111 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-log-httpd\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.724894 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.724978 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-config-data\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.725028 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-run-httpd\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.725053 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-log-httpd\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.725132 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkdx2\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-kube-api-access-kkdx2\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: E1004 07:28:05.725212 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:28:05 crc kubenswrapper[4987]: E1004 07:28:05.725261 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9: configmap "swift-ring-files" not found Oct 04 07:28:05 crc kubenswrapper[4987]: E1004 07:28:05.725395 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift podName:a7c67d2f-aa98-45e4-8dc4-ba02f6051378 nodeName:}" failed. No retries permitted until 2025-10-04 07:28:06.225359746 +0000 UTC m=+1103.280258598 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift") pod "swift-proxy-59cb459c9f-qhgk9" (UID: "a7c67d2f-aa98-45e4-8dc4-ba02f6051378") : configmap "swift-ring-files" not found Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.725778 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-run-httpd\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.725923 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-log-httpd\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.740645 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-config-data\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:05 crc kubenswrapper[4987]: I1004 07:28:05.743931 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkdx2\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-kube-api-access-kkdx2\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:06 crc kubenswrapper[4987]: I1004 07:28:06.236264 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:06 crc kubenswrapper[4987]: E1004 07:28:06.236498 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:28:06 crc kubenswrapper[4987]: E1004 07:28:06.236534 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9: configmap "swift-ring-files" not found Oct 04 07:28:06 crc kubenswrapper[4987]: E1004 07:28:06.236608 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift podName:a7c67d2f-aa98-45e4-8dc4-ba02f6051378 nodeName:}" failed. No retries permitted until 2025-10-04 07:28:07.236585508 +0000 UTC m=+1104.291484350 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift") pod "swift-proxy-59cb459c9f-qhgk9" (UID: "a7c67d2f-aa98-45e4-8dc4-ba02f6051378") : configmap "swift-ring-files" not found Oct 04 07:28:06 crc kubenswrapper[4987]: I1004 07:28:06.420382 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:28:06 crc kubenswrapper[4987]: I1004 07:28:06.420864 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:28:06 crc kubenswrapper[4987]: I1004 07:28:06.450669 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:28:06 crc kubenswrapper[4987]: I1004 07:28:06.608700 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-f5msg" Oct 04 07:28:07 crc kubenswrapper[4987]: I1004 07:28:07.248330 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:07 crc kubenswrapper[4987]: E1004 07:28:07.248658 4987 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Oct 04 07:28:07 crc kubenswrapper[4987]: E1004 07:28:07.248801 4987 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9: configmap "swift-ring-files" not found Oct 04 07:28:07 crc kubenswrapper[4987]: E1004 07:28:07.248916 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift podName:a7c67d2f-aa98-45e4-8dc4-ba02f6051378 nodeName:}" failed. No retries permitted until 2025-10-04 07:28:09.248881185 +0000 UTC m=+1106.303780037 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift") pod "swift-proxy-59cb459c9f-qhgk9" (UID: "a7c67d2f-aa98-45e4-8dc4-ba02f6051378") : configmap "swift-ring-files" not found Oct 04 07:28:08 crc kubenswrapper[4987]: I1004 07:28:08.594481 4987 generic.go:334] "Generic (PLEG): container finished" podID="a531f5c9-a54b-4016-83d4-57e4ced832d6" containerID="345d5d44dbc11c3bafc6bb41166f7fe2b99c97df85013c5d3cc109bf4216450a" exitCode=0 Oct 04 07:28:08 crc kubenswrapper[4987]: I1004 07:28:08.594601 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" event={"ID":"a531f5c9-a54b-4016-83d4-57e4ced832d6","Type":"ContainerDied","Data":"345d5d44dbc11c3bafc6bb41166f7fe2b99c97df85013c5d3cc109bf4216450a"} Oct 04 07:28:08 crc kubenswrapper[4987]: I1004 07:28:08.932271 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq"] Oct 04 07:28:08 crc kubenswrapper[4987]: I1004 07:28:08.933535 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:08 crc kubenswrapper[4987]: I1004 07:28:08.935662 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hwsts" Oct 04 07:28:08 crc kubenswrapper[4987]: I1004 07:28:08.947295 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq"] Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.078754 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-259rr\" (UniqueName: \"kubernetes.io/projected/fd50d7a6-1e67-4764-b700-39081ed2da4a-kube-api-access-259rr\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.078833 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-bundle\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.078956 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-util\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.180547 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-259rr\" (UniqueName: \"kubernetes.io/projected/fd50d7a6-1e67-4764-b700-39081ed2da4a-kube-api-access-259rr\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.180659 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-bundle\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.180837 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-util\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.181504 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-util\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.182027 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-bundle\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.211426 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-259rr\" (UniqueName: \"kubernetes.io/projected/fd50d7a6-1e67-4764-b700-39081ed2da4a-kube-api-access-259rr\") pod \"39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.251786 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.282566 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.287540 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7c67d2f-aa98-45e4-8dc4-ba02f6051378-etc-swift\") pod \"swift-proxy-59cb459c9f-qhgk9\" (UID: \"a7c67d2f-aa98-45e4-8dc4-ba02f6051378\") " pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.378943 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.691253 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq"] Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.853243 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9"] Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.853270 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:28:09 crc kubenswrapper[4987]: W1004 07:28:09.856019 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7c67d2f_aa98_45e4_8dc4_ba02f6051378.slice/crio-699374ba87912850095640f903d20f17a192f964b5153512f80aabab1bae1488 WatchSource:0}: Error finding container 699374ba87912850095640f903d20f17a192f964b5153512f80aabab1bae1488: Status 404 returned error can't find the container with id 699374ba87912850095640f903d20f17a192f964b5153512f80aabab1bae1488 Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.997381 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a531f5c9-a54b-4016-83d4-57e4ced832d6-etc-swift\") pod \"a531f5c9-a54b-4016-83d4-57e4ced832d6\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.997810 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-dispersionconf\") pod \"a531f5c9-a54b-4016-83d4-57e4ced832d6\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.997845 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-scripts\") pod \"a531f5c9-a54b-4016-83d4-57e4ced832d6\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.997867 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z28wq\" (UniqueName: \"kubernetes.io/projected/a531f5c9-a54b-4016-83d4-57e4ced832d6-kube-api-access-z28wq\") pod \"a531f5c9-a54b-4016-83d4-57e4ced832d6\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.997975 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-swiftconf\") pod \"a531f5c9-a54b-4016-83d4-57e4ced832d6\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.997997 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-ring-data-devices\") pod \"a531f5c9-a54b-4016-83d4-57e4ced832d6\" (UID: \"a531f5c9-a54b-4016-83d4-57e4ced832d6\") " Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.998959 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a531f5c9-a54b-4016-83d4-57e4ced832d6" (UID: "a531f5c9-a54b-4016-83d4-57e4ced832d6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:28:09 crc kubenswrapper[4987]: I1004 07:28:09.998976 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a531f5c9-a54b-4016-83d4-57e4ced832d6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a531f5c9-a54b-4016-83d4-57e4ced832d6" (UID: "a531f5c9-a54b-4016-83d4-57e4ced832d6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.002150 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a531f5c9-a54b-4016-83d4-57e4ced832d6-kube-api-access-z28wq" (OuterVolumeSpecName: "kube-api-access-z28wq") pod "a531f5c9-a54b-4016-83d4-57e4ced832d6" (UID: "a531f5c9-a54b-4016-83d4-57e4ced832d6"). InnerVolumeSpecName "kube-api-access-z28wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.003992 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a531f5c9-a54b-4016-83d4-57e4ced832d6" (UID: "a531f5c9-a54b-4016-83d4-57e4ced832d6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.014422 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-scripts" (OuterVolumeSpecName: "scripts") pod "a531f5c9-a54b-4016-83d4-57e4ced832d6" (UID: "a531f5c9-a54b-4016-83d4-57e4ced832d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.015848 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a531f5c9-a54b-4016-83d4-57e4ced832d6" (UID: "a531f5c9-a54b-4016-83d4-57e4ced832d6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.102422 4987 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.102525 4987 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a531f5c9-a54b-4016-83d4-57e4ced832d6-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.102550 4987 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.102602 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a531f5c9-a54b-4016-83d4-57e4ced832d6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.102674 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z28wq\" (UniqueName: \"kubernetes.io/projected/a531f5c9-a54b-4016-83d4-57e4ced832d6-kube-api-access-z28wq\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.102699 4987 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a531f5c9-a54b-4016-83d4-57e4ced832d6-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.609289 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" event={"ID":"a531f5c9-a54b-4016-83d4-57e4ced832d6","Type":"ContainerDied","Data":"086653578db9860abb8f523ca03f70b8c3e7b0028e6ee045deb7a071a1981522"} Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.609333 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="086653578db9860abb8f523ca03f70b8c3e7b0028e6ee045deb7a071a1981522" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.609397 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-zj4nq" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.611532 4987 generic.go:334] "Generic (PLEG): container finished" podID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerID="2bd678ea259f8e4b1c313e5e5afd9c770f35a73bae00949cc6950354d21adf48" exitCode=0 Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.611587 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" event={"ID":"fd50d7a6-1e67-4764-b700-39081ed2da4a","Type":"ContainerDied","Data":"2bd678ea259f8e4b1c313e5e5afd9c770f35a73bae00949cc6950354d21adf48"} Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.611604 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" event={"ID":"fd50d7a6-1e67-4764-b700-39081ed2da4a","Type":"ContainerStarted","Data":"ef32ffa726c6d99303aae322a6eae606210709f351920c6fe9fd6ea29dce0125"} Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.614960 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" event={"ID":"a7c67d2f-aa98-45e4-8dc4-ba02f6051378","Type":"ContainerStarted","Data":"bc704551f7f58e13dc97900ae1956d509d51dc55893970acb6f83d43d02c186a"} Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.614985 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" event={"ID":"a7c67d2f-aa98-45e4-8dc4-ba02f6051378","Type":"ContainerStarted","Data":"5c96e1c3347987d60ce677272523e273b6e29233e56ca03494dc27fbddb929d9"} Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.615505 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" event={"ID":"a7c67d2f-aa98-45e4-8dc4-ba02f6051378","Type":"ContainerStarted","Data":"699374ba87912850095640f903d20f17a192f964b5153512f80aabab1bae1488"} Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.615522 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.615532 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:10 crc kubenswrapper[4987]: I1004 07:28:10.658304 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" podStartSLOduration=5.6582877 podStartE2EDuration="5.6582877s" podCreationTimestamp="2025-10-04 07:28:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:28:10.651160115 +0000 UTC m=+1107.706058957" watchObservedRunningTime="2025-10-04 07:28:10.6582877 +0000 UTC m=+1107.713186542" Oct 04 07:28:11 crc kubenswrapper[4987]: I1004 07:28:11.622776 4987 generic.go:334] "Generic (PLEG): container finished" podID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerID="a3482df7589ac9f0a8ee3571b3e18b415506b8af31d51e1caf867818671cc425" exitCode=0 Oct 04 07:28:11 crc kubenswrapper[4987]: I1004 07:28:11.622828 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" event={"ID":"fd50d7a6-1e67-4764-b700-39081ed2da4a","Type":"ContainerDied","Data":"a3482df7589ac9f0a8ee3571b3e18b415506b8af31d51e1caf867818671cc425"} Oct 04 07:28:12 crc kubenswrapper[4987]: I1004 07:28:12.631708 4987 generic.go:334] "Generic (PLEG): container finished" podID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerID="eca0b8c33c067e360fda2ce1160096a7bb88f8bddb0aee028dfbfd0f39ab3670" exitCode=0 Oct 04 07:28:12 crc kubenswrapper[4987]: I1004 07:28:12.631819 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" event={"ID":"fd50d7a6-1e67-4764-b700-39081ed2da4a","Type":"ContainerDied","Data":"eca0b8c33c067e360fda2ce1160096a7bb88f8bddb0aee028dfbfd0f39ab3670"} Oct 04 07:28:12 crc kubenswrapper[4987]: I1004 07:28:12.702955 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:28:12 crc kubenswrapper[4987]: I1004 07:28:12.703024 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:28:13 crc kubenswrapper[4987]: I1004 07:28:13.968207 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.062290 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-259rr\" (UniqueName: \"kubernetes.io/projected/fd50d7a6-1e67-4764-b700-39081ed2da4a-kube-api-access-259rr\") pod \"fd50d7a6-1e67-4764-b700-39081ed2da4a\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.062655 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-util\") pod \"fd50d7a6-1e67-4764-b700-39081ed2da4a\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.062721 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-bundle\") pod \"fd50d7a6-1e67-4764-b700-39081ed2da4a\" (UID: \"fd50d7a6-1e67-4764-b700-39081ed2da4a\") " Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.063828 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-bundle" (OuterVolumeSpecName: "bundle") pod "fd50d7a6-1e67-4764-b700-39081ed2da4a" (UID: "fd50d7a6-1e67-4764-b700-39081ed2da4a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.071692 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd50d7a6-1e67-4764-b700-39081ed2da4a-kube-api-access-259rr" (OuterVolumeSpecName: "kube-api-access-259rr") pod "fd50d7a6-1e67-4764-b700-39081ed2da4a" (UID: "fd50d7a6-1e67-4764-b700-39081ed2da4a"). InnerVolumeSpecName "kube-api-access-259rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.076812 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-util" (OuterVolumeSpecName: "util") pod "fd50d7a6-1e67-4764-b700-39081ed2da4a" (UID: "fd50d7a6-1e67-4764-b700-39081ed2da4a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.164967 4987 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-util\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.165017 4987 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fd50d7a6-1e67-4764-b700-39081ed2da4a-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.165031 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-259rr\" (UniqueName: \"kubernetes.io/projected/fd50d7a6-1e67-4764-b700-39081ed2da4a-kube-api-access-259rr\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.648383 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" event={"ID":"fd50d7a6-1e67-4764-b700-39081ed2da4a","Type":"ContainerDied","Data":"ef32ffa726c6d99303aae322a6eae606210709f351920c6fe9fd6ea29dce0125"} Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.648431 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef32ffa726c6d99303aae322a6eae606210709f351920c6fe9fd6ea29dce0125" Oct 04 07:28:14 crc kubenswrapper[4987]: I1004 07:28:14.648465 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq" Oct 04 07:28:19 crc kubenswrapper[4987]: I1004 07:28:19.382184 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:19 crc kubenswrapper[4987]: I1004 07:28:19.386199 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-59cb459c9f-qhgk9" Oct 04 07:28:20 crc kubenswrapper[4987]: I1004 07:28:20.671907 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:28:20 crc kubenswrapper[4987]: I1004 07:28:20.684369 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/70e18ecc-8da8-4423-8eff-39f07bf2383f-etc-swift\") pod \"swift-storage-0\" (UID: \"70e18ecc-8da8-4423-8eff-39f07bf2383f\") " pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:28:20 crc kubenswrapper[4987]: I1004 07:28:20.843409 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Oct 04 07:28:21 crc kubenswrapper[4987]: I1004 07:28:21.336898 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Oct 04 07:28:21 crc kubenswrapper[4987]: I1004 07:28:21.718506 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"2871c6025ab5af217151330782ca0e946e3dca735672ec4471fb768c78b7842f"} Oct 04 07:28:22 crc kubenswrapper[4987]: I1004 07:28:22.732677 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"db6f5f26567619e2c79453b3b40b7894b4642369502a62874bbb4844a14a447f"} Oct 04 07:28:23 crc kubenswrapper[4987]: I1004 07:28:23.766698 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"09e42ca503b2226102f9489e147ed22f1b52327e6ebcda3e557909f41759e014"} Oct 04 07:28:23 crc kubenswrapper[4987]: I1004 07:28:23.767259 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"f6f5cc1153f0f0060afec3aec85de524a22563755510f8e40af6e46477561373"} Oct 04 07:28:23 crc kubenswrapper[4987]: I1004 07:28:23.767273 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"694b12000b80bd26e5229223f05d3cc103cb288c98df3b2273d49447c38d250b"} Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.492450 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc"] Oct 04 07:28:24 crc kubenswrapper[4987]: E1004 07:28:24.493233 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="pull" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.493250 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="pull" Oct 04 07:28:24 crc kubenswrapper[4987]: E1004 07:28:24.493276 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="extract" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.493285 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="extract" Oct 04 07:28:24 crc kubenswrapper[4987]: E1004 07:28:24.493297 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a531f5c9-a54b-4016-83d4-57e4ced832d6" containerName="swift-ring-rebalance" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.493306 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="a531f5c9-a54b-4016-83d4-57e4ced832d6" containerName="swift-ring-rebalance" Oct 04 07:28:24 crc kubenswrapper[4987]: E1004 07:28:24.493317 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="util" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.493323 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="util" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.493460 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="a531f5c9-a54b-4016-83d4-57e4ced832d6" containerName="swift-ring-rebalance" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.493500 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd50d7a6-1e67-4764-b700-39081ed2da4a" containerName="extract" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.494321 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.496357 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.498148 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-68nl9" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.510168 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc"] Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.654162 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btr4n\" (UniqueName: \"kubernetes.io/projected/4f9c4992-f741-41e4-9f30-90c4d1c165a5-kube-api-access-btr4n\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.654722 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4f9c4992-f741-41e4-9f30-90c4d1c165a5-webhook-cert\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.654904 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4f9c4992-f741-41e4-9f30-90c4d1c165a5-apiservice-cert\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.756822 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btr4n\" (UniqueName: \"kubernetes.io/projected/4f9c4992-f741-41e4-9f30-90c4d1c165a5-kube-api-access-btr4n\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.756978 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4f9c4992-f741-41e4-9f30-90c4d1c165a5-webhook-cert\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.757018 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4f9c4992-f741-41e4-9f30-90c4d1c165a5-apiservice-cert\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.765323 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4f9c4992-f741-41e4-9f30-90c4d1c165a5-webhook-cert\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.776072 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btr4n\" (UniqueName: \"kubernetes.io/projected/4f9c4992-f741-41e4-9f30-90c4d1c165a5-kube-api-access-btr4n\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.776574 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4f9c4992-f741-41e4-9f30-90c4d1c165a5-apiservice-cert\") pod \"glance-operator-controller-manager-6468bd8bb6-p22sc\" (UID: \"4f9c4992-f741-41e4-9f30-90c4d1c165a5\") " pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.793381 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"1a5b6d0585a6d10c732f153c32e1ba1ae044e8d1ceba6d06a3107f5cd69cedff"} Oct 04 07:28:24 crc kubenswrapper[4987]: I1004 07:28:24.822788 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:25 crc kubenswrapper[4987]: I1004 07:28:25.293512 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc"] Oct 04 07:28:25 crc kubenswrapper[4987]: I1004 07:28:25.815328 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"35be0fe6632db14414a56840565bff83650f469c6f26bbf890aa44c4bf53599d"} Oct 04 07:28:25 crc kubenswrapper[4987]: I1004 07:28:25.815408 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"8566ca7d5f8bcf57e86707d99ba5a33d0f666e8e41a93a3b90fb0f3707d2746d"} Oct 04 07:28:25 crc kubenswrapper[4987]: I1004 07:28:25.815422 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"5cc7cbc6b372d9e49be721ad4873ef7aaf3839039bc7d4e6833b2f4e9ec4b869"} Oct 04 07:28:25 crc kubenswrapper[4987]: I1004 07:28:25.816996 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" event={"ID":"4f9c4992-f741-41e4-9f30-90c4d1c165a5","Type":"ContainerStarted","Data":"0c5f83c59d4dd93afb30d91939c34ae6b68c81da0f31f8e0f11fccba74561a89"} Oct 04 07:28:27 crc kubenswrapper[4987]: I1004 07:28:27.839436 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" event={"ID":"4f9c4992-f741-41e4-9f30-90c4d1c165a5","Type":"ContainerStarted","Data":"d0e72a8a675831244ec047ff748695f4c4b76d15c7ff0db22c4447e98446ac5f"} Oct 04 07:28:27 crc kubenswrapper[4987]: I1004 07:28:27.846496 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"66f2d707cbaa55d32227abe59b007bb09c973d22d57777cab631806c769b9136"} Oct 04 07:28:27 crc kubenswrapper[4987]: I1004 07:28:27.846546 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"0bbd967301896fd14f5b490ac776960f4080145b399f3fd61d11759ae49ce2ec"} Oct 04 07:28:27 crc kubenswrapper[4987]: I1004 07:28:27.846555 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"3ee88e7fe324c071e04b0eb1ae20cde2e93988e35438e609330f22da50d65c1a"} Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.869676 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"875f1c4392aa5ae59f6690a136323d9b314ac17de56da48ec5ffad6c228318ef"} Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.870081 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"6a5e6392418600db6df16e7fc58376dce09f177eec6e498ebcc33a8fe408a135"} Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.870094 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"b67cf7af4e9ba43815278f8b92b88e35fefbcab53251dd22fd71c729043d16d1"} Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.870492 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"70e18ecc-8da8-4423-8eff-39f07bf2383f","Type":"ContainerStarted","Data":"1e50d680344f8622992a4ca33cd3014bbe808573dbb9202980b5a4a7828ce291"} Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.877452 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" event={"ID":"4f9c4992-f741-41e4-9f30-90c4d1c165a5","Type":"ContainerStarted","Data":"8358ab34047987a1a6f8d68eca128b88821c56dec5feea0ca93ae74ec3d64040"} Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.877704 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.972944 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" podStartSLOduration=2.355638351 podStartE2EDuration="4.972912011s" podCreationTimestamp="2025-10-04 07:28:24 +0000 UTC" firstStartedPulling="2025-10-04 07:28:25.300128653 +0000 UTC m=+1122.355027495" lastFinishedPulling="2025-10-04 07:28:27.917402313 +0000 UTC m=+1124.972301155" observedRunningTime="2025-10-04 07:28:28.971098254 +0000 UTC m=+1126.025997186" watchObservedRunningTime="2025-10-04 07:28:28.972912011 +0000 UTC m=+1126.027810883" Oct 04 07:28:28 crc kubenswrapper[4987]: I1004 07:28:28.978929 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=36.069019093 podStartE2EDuration="41.97889864s" podCreationTimestamp="2025-10-04 07:27:47 +0000 UTC" firstStartedPulling="2025-10-04 07:28:21.34841055 +0000 UTC m=+1118.403309412" lastFinishedPulling="2025-10-04 07:28:27.258290077 +0000 UTC m=+1124.313188959" observedRunningTime="2025-10-04 07:28:28.933581443 +0000 UTC m=+1125.988480295" watchObservedRunningTime="2025-10-04 07:28:28.97889864 +0000 UTC m=+1126.033797522" Oct 04 07:28:34 crc kubenswrapper[4987]: I1004 07:28:34.829010 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-6468bd8bb6-p22sc" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.066566 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.068432 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.071680 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.073678 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.073934 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.075689 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-sfbrn" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.090679 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.118878 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-qzmzd"] Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.127727 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.130565 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-qzmzd"] Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.180045 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcpgk\" (UniqueName: \"kubernetes.io/projected/d1343ba4-662f-4e15-a437-fb80f6987c0e-kube-api-access-pcpgk\") pod \"glance-db-create-qzmzd\" (UID: \"d1343ba4-662f-4e15-a437-fb80f6987c0e\") " pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.180112 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-scripts\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.180149 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.180171 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config-secret\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.180194 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nxkv\" (UniqueName: \"kubernetes.io/projected/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-kube-api-access-7nxkv\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.281709 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcpgk\" (UniqueName: \"kubernetes.io/projected/d1343ba4-662f-4e15-a437-fb80f6987c0e-kube-api-access-pcpgk\") pod \"glance-db-create-qzmzd\" (UID: \"d1343ba4-662f-4e15-a437-fb80f6987c0e\") " pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.281820 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-scripts\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.281904 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.281941 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config-secret\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.281975 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nxkv\" (UniqueName: \"kubernetes.io/projected/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-kube-api-access-7nxkv\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.283459 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.283511 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-scripts\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.289179 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config-secret\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.299159 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nxkv\" (UniqueName: \"kubernetes.io/projected/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-kube-api-access-7nxkv\") pod \"openstackclient\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.314461 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcpgk\" (UniqueName: \"kubernetes.io/projected/d1343ba4-662f-4e15-a437-fb80f6987c0e-kube-api-access-pcpgk\") pod \"glance-db-create-qzmzd\" (UID: \"d1343ba4-662f-4e15-a437-fb80f6987c0e\") " pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.403365 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.456015 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.696872 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:28:38 crc kubenswrapper[4987]: W1004 07:28:38.701361 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd83077b_d8ac_4e5c_9a9d_2894d32964d6.slice/crio-c4763304cc3d816a1402d22b4f598663d530c3d899f792d30088361eb6344f21 WatchSource:0}: Error finding container c4763304cc3d816a1402d22b4f598663d530c3d899f792d30088361eb6344f21: Status 404 returned error can't find the container with id c4763304cc3d816a1402d22b4f598663d530c3d899f792d30088361eb6344f21 Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.754996 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-qzmzd"] Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.987103 4987 generic.go:334] "Generic (PLEG): container finished" podID="d1343ba4-662f-4e15-a437-fb80f6987c0e" containerID="6b8ad1c4bf6f6f36ac8f306829df4bf04e058c17c03b08d77d58b06236523b6a" exitCode=0 Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.987220 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qzmzd" event={"ID":"d1343ba4-662f-4e15-a437-fb80f6987c0e","Type":"ContainerDied","Data":"6b8ad1c4bf6f6f36ac8f306829df4bf04e058c17c03b08d77d58b06236523b6a"} Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.987266 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qzmzd" event={"ID":"d1343ba4-662f-4e15-a437-fb80f6987c0e","Type":"ContainerStarted","Data":"9e55963e2a79422dd5005c6985c51a3e058f5a60a9fb6f8a0c19962d80029e6e"} Oct 04 07:28:38 crc kubenswrapper[4987]: I1004 07:28:38.989114 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"fd83077b-d8ac-4e5c-9a9d-2894d32964d6","Type":"ContainerStarted","Data":"c4763304cc3d816a1402d22b4f598663d530c3d899f792d30088361eb6344f21"} Oct 04 07:28:40 crc kubenswrapper[4987]: I1004 07:28:40.295723 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:40 crc kubenswrapper[4987]: I1004 07:28:40.414433 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcpgk\" (UniqueName: \"kubernetes.io/projected/d1343ba4-662f-4e15-a437-fb80f6987c0e-kube-api-access-pcpgk\") pod \"d1343ba4-662f-4e15-a437-fb80f6987c0e\" (UID: \"d1343ba4-662f-4e15-a437-fb80f6987c0e\") " Oct 04 07:28:40 crc kubenswrapper[4987]: I1004 07:28:40.425815 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1343ba4-662f-4e15-a437-fb80f6987c0e-kube-api-access-pcpgk" (OuterVolumeSpecName: "kube-api-access-pcpgk") pod "d1343ba4-662f-4e15-a437-fb80f6987c0e" (UID: "d1343ba4-662f-4e15-a437-fb80f6987c0e"). InnerVolumeSpecName "kube-api-access-pcpgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:28:40 crc kubenswrapper[4987]: I1004 07:28:40.517269 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcpgk\" (UniqueName: \"kubernetes.io/projected/d1343ba4-662f-4e15-a437-fb80f6987c0e-kube-api-access-pcpgk\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:41 crc kubenswrapper[4987]: I1004 07:28:41.005144 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-qzmzd" event={"ID":"d1343ba4-662f-4e15-a437-fb80f6987c0e","Type":"ContainerDied","Data":"9e55963e2a79422dd5005c6985c51a3e058f5a60a9fb6f8a0c19962d80029e6e"} Oct 04 07:28:41 crc kubenswrapper[4987]: I1004 07:28:41.005186 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e55963e2a79422dd5005c6985c51a3e058f5a60a9fb6f8a0c19962d80029e6e" Oct 04 07:28:41 crc kubenswrapper[4987]: I1004 07:28:41.005232 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-qzmzd" Oct 04 07:28:42 crc kubenswrapper[4987]: I1004 07:28:42.703119 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:28:42 crc kubenswrapper[4987]: I1004 07:28:42.703441 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.086534 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"fd83077b-d8ac-4e5c-9a9d-2894d32964d6","Type":"ContainerStarted","Data":"dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0"} Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.100559 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-f0dc-account-create-kskcp"] Oct 04 07:28:48 crc kubenswrapper[4987]: E1004 07:28:48.100890 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1343ba4-662f-4e15-a437-fb80f6987c0e" containerName="mariadb-database-create" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.100903 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1343ba4-662f-4e15-a437-fb80f6987c0e" containerName="mariadb-database-create" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.101040 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1343ba4-662f-4e15-a437-fb80f6987c0e" containerName="mariadb-database-create" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.101526 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.107460 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.112994 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f0dc-account-create-kskcp"] Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.128039 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.611564462 podStartE2EDuration="10.128017922s" podCreationTimestamp="2025-10-04 07:28:38 +0000 UTC" firstStartedPulling="2025-10-04 07:28:38.704763346 +0000 UTC m=+1135.759662188" lastFinishedPulling="2025-10-04 07:28:47.221216806 +0000 UTC m=+1144.276115648" observedRunningTime="2025-10-04 07:28:48.116675625 +0000 UTC m=+1145.171574467" watchObservedRunningTime="2025-10-04 07:28:48.128017922 +0000 UTC m=+1145.182916764" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.154023 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxwg8\" (UniqueName: \"kubernetes.io/projected/4b53ffe9-da04-4a9f-bf00-557c359928f2-kube-api-access-mxwg8\") pod \"glance-f0dc-account-create-kskcp\" (UID: \"4b53ffe9-da04-4a9f-bf00-557c359928f2\") " pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.255411 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxwg8\" (UniqueName: \"kubernetes.io/projected/4b53ffe9-da04-4a9f-bf00-557c359928f2-kube-api-access-mxwg8\") pod \"glance-f0dc-account-create-kskcp\" (UID: \"4b53ffe9-da04-4a9f-bf00-557c359928f2\") " pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.274515 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxwg8\" (UniqueName: \"kubernetes.io/projected/4b53ffe9-da04-4a9f-bf00-557c359928f2-kube-api-access-mxwg8\") pod \"glance-f0dc-account-create-kskcp\" (UID: \"4b53ffe9-da04-4a9f-bf00-557c359928f2\") " pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.425005 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:48 crc kubenswrapper[4987]: I1004 07:28:48.854447 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f0dc-account-create-kskcp"] Oct 04 07:28:48 crc kubenswrapper[4987]: W1004 07:28:48.861115 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b53ffe9_da04_4a9f_bf00_557c359928f2.slice/crio-5af44163c010dbec98cb7ffe3fc4827ce655974f6c7ba1de9e7b5671798d8b97 WatchSource:0}: Error finding container 5af44163c010dbec98cb7ffe3fc4827ce655974f6c7ba1de9e7b5671798d8b97: Status 404 returned error can't find the container with id 5af44163c010dbec98cb7ffe3fc4827ce655974f6c7ba1de9e7b5671798d8b97 Oct 04 07:28:49 crc kubenswrapper[4987]: I1004 07:28:49.095377 4987 generic.go:334] "Generic (PLEG): container finished" podID="4b53ffe9-da04-4a9f-bf00-557c359928f2" containerID="a188687e0e4e09e1ccc47d4d9c2d7d26742ec18517d2ccc0286fe72ec878db10" exitCode=0 Oct 04 07:28:49 crc kubenswrapper[4987]: I1004 07:28:49.096260 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" event={"ID":"4b53ffe9-da04-4a9f-bf00-557c359928f2","Type":"ContainerDied","Data":"a188687e0e4e09e1ccc47d4d9c2d7d26742ec18517d2ccc0286fe72ec878db10"} Oct 04 07:28:49 crc kubenswrapper[4987]: I1004 07:28:49.096297 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" event={"ID":"4b53ffe9-da04-4a9f-bf00-557c359928f2","Type":"ContainerStarted","Data":"5af44163c010dbec98cb7ffe3fc4827ce655974f6c7ba1de9e7b5671798d8b97"} Oct 04 07:28:50 crc kubenswrapper[4987]: I1004 07:28:50.370979 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:50 crc kubenswrapper[4987]: I1004 07:28:50.487925 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxwg8\" (UniqueName: \"kubernetes.io/projected/4b53ffe9-da04-4a9f-bf00-557c359928f2-kube-api-access-mxwg8\") pod \"4b53ffe9-da04-4a9f-bf00-557c359928f2\" (UID: \"4b53ffe9-da04-4a9f-bf00-557c359928f2\") " Oct 04 07:28:50 crc kubenswrapper[4987]: I1004 07:28:50.499108 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b53ffe9-da04-4a9f-bf00-557c359928f2-kube-api-access-mxwg8" (OuterVolumeSpecName: "kube-api-access-mxwg8") pod "4b53ffe9-da04-4a9f-bf00-557c359928f2" (UID: "4b53ffe9-da04-4a9f-bf00-557c359928f2"). InnerVolumeSpecName "kube-api-access-mxwg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:28:50 crc kubenswrapper[4987]: I1004 07:28:50.590173 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxwg8\" (UniqueName: \"kubernetes.io/projected/4b53ffe9-da04-4a9f-bf00-557c359928f2-kube-api-access-mxwg8\") on node \"crc\" DevicePath \"\"" Oct 04 07:28:51 crc kubenswrapper[4987]: I1004 07:28:51.113303 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" event={"ID":"4b53ffe9-da04-4a9f-bf00-557c359928f2","Type":"ContainerDied","Data":"5af44163c010dbec98cb7ffe3fc4827ce655974f6c7ba1de9e7b5671798d8b97"} Oct 04 07:28:51 crc kubenswrapper[4987]: I1004 07:28:51.113349 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5af44163c010dbec98cb7ffe3fc4827ce655974f6c7ba1de9e7b5671798d8b97" Oct 04 07:28:51 crc kubenswrapper[4987]: I1004 07:28:51.113371 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f0dc-account-create-kskcp" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.141730 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-nmfp4"] Oct 04 07:28:53 crc kubenswrapper[4987]: E1004 07:28:53.142067 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b53ffe9-da04-4a9f-bf00-557c359928f2" containerName="mariadb-account-create" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.142084 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b53ffe9-da04-4a9f-bf00-557c359928f2" containerName="mariadb-account-create" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.142265 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b53ffe9-da04-4a9f-bf00-557c359928f2" containerName="mariadb-account-create" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.142806 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.145049 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-rfng2" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.145078 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.155415 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nmfp4"] Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.233710 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-config-data\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.233795 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-db-sync-config-data\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.233819 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp2rn\" (UniqueName: \"kubernetes.io/projected/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-kube-api-access-kp2rn\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.335104 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-config-data\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.335499 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-db-sync-config-data\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.335523 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp2rn\" (UniqueName: \"kubernetes.io/projected/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-kube-api-access-kp2rn\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.343991 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-db-sync-config-data\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.344451 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-config-data\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.365352 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp2rn\" (UniqueName: \"kubernetes.io/projected/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-kube-api-access-kp2rn\") pod \"glance-db-sync-nmfp4\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.459197 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:28:53 crc kubenswrapper[4987]: I1004 07:28:53.916779 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nmfp4"] Oct 04 07:28:54 crc kubenswrapper[4987]: I1004 07:28:54.133990 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nmfp4" event={"ID":"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f","Type":"ContainerStarted","Data":"4a6c8f5591dcdd41e63ed404bb0207bb0239ee860927a641ee2b69c691cd45cc"} Oct 04 07:29:06 crc kubenswrapper[4987]: I1004 07:29:06.243794 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nmfp4" event={"ID":"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f","Type":"ContainerStarted","Data":"fc10285d1d547f376c0cf0fce7788cb81b08854e00c7b5cbe8869d391fe4ee71"} Oct 04 07:29:06 crc kubenswrapper[4987]: I1004 07:29:06.267046 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-nmfp4" podStartSLOduration=1.8099799619999999 podStartE2EDuration="13.267019334s" podCreationTimestamp="2025-10-04 07:28:53 +0000 UTC" firstStartedPulling="2025-10-04 07:28:53.926334966 +0000 UTC m=+1150.981233818" lastFinishedPulling="2025-10-04 07:29:05.383374348 +0000 UTC m=+1162.438273190" observedRunningTime="2025-10-04 07:29:06.266415345 +0000 UTC m=+1163.321314197" watchObservedRunningTime="2025-10-04 07:29:06.267019334 +0000 UTC m=+1163.321918176" Oct 04 07:29:12 crc kubenswrapper[4987]: I1004 07:29:12.703371 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:29:12 crc kubenswrapper[4987]: I1004 07:29:12.704125 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:29:12 crc kubenswrapper[4987]: I1004 07:29:12.704205 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:29:12 crc kubenswrapper[4987]: I1004 07:29:12.705050 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9851ab3918efc181099219aac8c4cb4de265cdf08bb6a6a2096a913ebd0e22d6"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:29:12 crc kubenswrapper[4987]: I1004 07:29:12.705100 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://9851ab3918efc181099219aac8c4cb4de265cdf08bb6a6a2096a913ebd0e22d6" gracePeriod=600 Oct 04 07:29:13 crc kubenswrapper[4987]: I1004 07:29:13.306172 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="9851ab3918efc181099219aac8c4cb4de265cdf08bb6a6a2096a913ebd0e22d6" exitCode=0 Oct 04 07:29:13 crc kubenswrapper[4987]: I1004 07:29:13.306335 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"9851ab3918efc181099219aac8c4cb4de265cdf08bb6a6a2096a913ebd0e22d6"} Oct 04 07:29:13 crc kubenswrapper[4987]: I1004 07:29:13.306709 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"901e84e5b40d7f1df5e525731897cdcc9e887088810c1590e56d6e773c746b71"} Oct 04 07:29:13 crc kubenswrapper[4987]: I1004 07:29:13.306744 4987 scope.go:117] "RemoveContainer" containerID="914ae0228b00a348da279a19d0c789018ccba172664058b29703314fa8fd5ab8" Oct 04 07:29:14 crc kubenswrapper[4987]: I1004 07:29:14.324400 4987 generic.go:334] "Generic (PLEG): container finished" podID="8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" containerID="fc10285d1d547f376c0cf0fce7788cb81b08854e00c7b5cbe8869d391fe4ee71" exitCode=0 Oct 04 07:29:14 crc kubenswrapper[4987]: I1004 07:29:14.324938 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nmfp4" event={"ID":"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f","Type":"ContainerDied","Data":"fc10285d1d547f376c0cf0fce7788cb81b08854e00c7b5cbe8869d391fe4ee71"} Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.645323 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.735205 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-config-data\") pod \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.735279 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-db-sync-config-data\") pod \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.735370 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp2rn\" (UniqueName: \"kubernetes.io/projected/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-kube-api-access-kp2rn\") pod \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\" (UID: \"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f\") " Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.756028 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-kube-api-access-kp2rn" (OuterVolumeSpecName: "kube-api-access-kp2rn") pod "8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" (UID: "8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f"). InnerVolumeSpecName "kube-api-access-kp2rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.775910 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" (UID: "8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.833534 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-config-data" (OuterVolumeSpecName: "config-data") pod "8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" (UID: "8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.837468 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.837494 4987 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:15 crc kubenswrapper[4987]: I1004 07:29:15.837507 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp2rn\" (UniqueName: \"kubernetes.io/projected/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f-kube-api-access-kp2rn\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.341935 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nmfp4" event={"ID":"8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f","Type":"ContainerDied","Data":"4a6c8f5591dcdd41e63ed404bb0207bb0239ee860927a641ee2b69c691cd45cc"} Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.342363 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6c8f5591dcdd41e63ed404bb0207bb0239ee860927a641ee2b69c691cd45cc" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.342001 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nmfp4" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.820433 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:16 crc kubenswrapper[4987]: E1004 07:29:16.821428 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" containerName="glance-db-sync" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.821456 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" containerName="glance-db-sync" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.821874 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" containerName="glance-db-sync" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.831538 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.836114 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-rfng2" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.836316 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.836762 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.838952 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.841022 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.845693 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.854915 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954649 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-config-data\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954711 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-dev\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954736 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-nvme\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954762 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-config-data\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954786 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954821 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-lib-modules\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.954848 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-scripts\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955211 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-lib-modules\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955304 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-sys\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955397 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-logs\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955424 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955532 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955571 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-sys\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955612 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955826 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-httpd-run\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955940 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-run\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.955972 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89hm9\" (UniqueName: \"kubernetes.io/projected/e05172a1-4714-44c6-b500-41eed762e8ba-kube-api-access-89hm9\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956015 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956040 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956234 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956287 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ggz7\" (UniqueName: \"kubernetes.io/projected/2066e87b-2788-47e8-ba09-4a8039b07e26-kube-api-access-8ggz7\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956343 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-httpd-run\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956364 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956392 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-dev\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956611 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-scripts\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956691 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-run\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956732 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:16 crc kubenswrapper[4987]: I1004 07:29:16.956753 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-logs\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058479 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058529 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ggz7\" (UniqueName: \"kubernetes.io/projected/2066e87b-2788-47e8-ba09-4a8039b07e26-kube-api-access-8ggz7\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058581 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-httpd-run\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058601 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058620 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-dev\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058648 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-scripts\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058664 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-run\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058680 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058696 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-logs\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058719 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-config-data\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058733 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-dev\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058748 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-nvme\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058768 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-config-data\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058786 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058829 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-scripts\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058848 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-lib-modules\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058865 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-lib-modules\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058889 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-sys\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058907 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-logs\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058924 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058948 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058968 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-sys\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.058984 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059004 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-httpd-run\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059021 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-run\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059036 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89hm9\" (UniqueName: \"kubernetes.io/projected/e05172a1-4714-44c6-b500-41eed762e8ba-kube-api-access-89hm9\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059053 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059068 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059136 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.059450 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.060107 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-httpd-run\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.060350 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.062845 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-lib-modules\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.062884 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-dev\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063014 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-dev\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063087 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-run\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063338 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-nvme\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063460 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-sys\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063524 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063845 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-logs\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.063461 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064206 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-lib-modules\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064288 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-run\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064339 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064350 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-logs\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064401 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-sys\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064473 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064649 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-nvme\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064676 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-httpd-run\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.064771 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.068683 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-scripts\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.069458 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-config-data\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.069810 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-config-data\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.079607 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-scripts\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.088110 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ggz7\" (UniqueName: \"kubernetes.io/projected/2066e87b-2788-47e8-ba09-4a8039b07e26-kube-api-access-8ggz7\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.092559 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89hm9\" (UniqueName: \"kubernetes.io/projected/e05172a1-4714-44c6-b500-41eed762e8ba-kube-api-access-89hm9\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.097438 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.100009 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.103861 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.111278 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.155622 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.167468 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.574403 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:17 crc kubenswrapper[4987]: I1004 07:29:17.729725 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:17 crc kubenswrapper[4987]: W1004 07:29:17.732401 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode05172a1_4714_44c6_b500_41eed762e8ba.slice/crio-7d79988230b17e18e23015cb0326e72cefc71d3ab3a466097c1dfd509f05640a WatchSource:0}: Error finding container 7d79988230b17e18e23015cb0326e72cefc71d3ab3a466097c1dfd509f05640a: Status 404 returned error can't find the container with id 7d79988230b17e18e23015cb0326e72cefc71d3ab3a466097c1dfd509f05640a Oct 04 07:29:18 crc kubenswrapper[4987]: I1004 07:29:18.364509 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"e05172a1-4714-44c6-b500-41eed762e8ba","Type":"ContainerStarted","Data":"7d79988230b17e18e23015cb0326e72cefc71d3ab3a466097c1dfd509f05640a"} Oct 04 07:29:18 crc kubenswrapper[4987]: I1004 07:29:18.366673 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2066e87b-2788-47e8-ba09-4a8039b07e26","Type":"ContainerStarted","Data":"180f4d2e2e1627c438b01a8726061ac14028a3d78dd56612243d13a5425e7e2f"} Oct 04 07:29:18 crc kubenswrapper[4987]: I1004 07:29:18.505006 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.395557 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"e05172a1-4714-44c6-b500-41eed762e8ba","Type":"ContainerStarted","Data":"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7"} Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.395972 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-log" containerID="cri-o://12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5" gracePeriod=30 Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.396660 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"e05172a1-4714-44c6-b500-41eed762e8ba","Type":"ContainerStarted","Data":"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5"} Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.396546 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-httpd" containerID="cri-o://2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7" gracePeriod=30 Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.400192 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2066e87b-2788-47e8-ba09-4a8039b07e26","Type":"ContainerStarted","Data":"f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef"} Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.400247 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2066e87b-2788-47e8-ba09-4a8039b07e26","Type":"ContainerStarted","Data":"6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb"} Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.425446 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=5.425428264 podStartE2EDuration="5.425428264s" podCreationTimestamp="2025-10-04 07:29:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:29:21.41860724 +0000 UTC m=+1178.473506092" watchObservedRunningTime="2025-10-04 07:29:21.425428264 +0000 UTC m=+1178.480327106" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.446659 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=5.446603432 podStartE2EDuration="5.446603432s" podCreationTimestamp="2025-10-04 07:29:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:29:21.443223055 +0000 UTC m=+1178.498121917" watchObservedRunningTime="2025-10-04 07:29:21.446603432 +0000 UTC m=+1178.501502304" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.817908 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939221 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-config-data\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939267 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939315 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-var-locks-brick\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939377 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-run\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939484 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-run" (OuterVolumeSpecName: "run") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939521 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.939556 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.940844 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-nvme\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.940892 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-lib-modules\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.940919 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-dev\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.940954 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-httpd-run\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.940991 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-scripts\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941045 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-logs\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941018 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941072 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941050 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-dev" (OuterVolumeSpecName: "dev") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941097 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941069 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-iscsi\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941227 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89hm9\" (UniqueName: \"kubernetes.io/projected/e05172a1-4714-44c6-b500-41eed762e8ba-kube-api-access-89hm9\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941293 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-sys\") pod \"e05172a1-4714-44c6-b500-41eed762e8ba\" (UID: \"e05172a1-4714-44c6-b500-41eed762e8ba\") " Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941393 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941455 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-logs" (OuterVolumeSpecName: "logs") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.941502 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-sys" (OuterVolumeSpecName: "sys") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942087 4987 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942106 4987 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942120 4987 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-dev\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942130 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942139 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e05172a1-4714-44c6-b500-41eed762e8ba-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942151 4987 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942162 4987 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-sys\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942173 4987 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.942182 4987 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e05172a1-4714-44c6-b500-41eed762e8ba-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.947087 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.947109 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e05172a1-4714-44c6-b500-41eed762e8ba-kube-api-access-89hm9" (OuterVolumeSpecName: "kube-api-access-89hm9") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "kube-api-access-89hm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.947172 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-scripts" (OuterVolumeSpecName: "scripts") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.947705 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:29:21 crc kubenswrapper[4987]: I1004 07:29:21.981351 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-config-data" (OuterVolumeSpecName: "config-data") pod "e05172a1-4714-44c6-b500-41eed762e8ba" (UID: "e05172a1-4714-44c6-b500-41eed762e8ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.043870 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.043973 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.043995 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.044013 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e05172a1-4714-44c6-b500-41eed762e8ba-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.044027 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89hm9\" (UniqueName: \"kubernetes.io/projected/e05172a1-4714-44c6-b500-41eed762e8ba-kube-api-access-89hm9\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.063565 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.075866 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.147389 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.150879 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.424094 4987 generic.go:334] "Generic (PLEG): container finished" podID="e05172a1-4714-44c6-b500-41eed762e8ba" containerID="2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7" exitCode=143 Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.424698 4987 generic.go:334] "Generic (PLEG): container finished" podID="e05172a1-4714-44c6-b500-41eed762e8ba" containerID="12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5" exitCode=143 Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.424334 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"e05172a1-4714-44c6-b500-41eed762e8ba","Type":"ContainerDied","Data":"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7"} Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.425101 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"e05172a1-4714-44c6-b500-41eed762e8ba","Type":"ContainerDied","Data":"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5"} Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.425143 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"e05172a1-4714-44c6-b500-41eed762e8ba","Type":"ContainerDied","Data":"7d79988230b17e18e23015cb0326e72cefc71d3ab3a466097c1dfd509f05640a"} Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.425171 4987 scope.go:117] "RemoveContainer" containerID="2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.424483 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.466876 4987 scope.go:117] "RemoveContainer" containerID="12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.467055 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.475181 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.517984 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:22 crc kubenswrapper[4987]: E1004 07:29:22.518713 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-httpd" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.518733 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-httpd" Oct 04 07:29:22 crc kubenswrapper[4987]: E1004 07:29:22.518760 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-log" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.518768 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-log" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.518897 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-log" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.518906 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" containerName="glance-httpd" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.519614 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.525326 4987 scope.go:117] "RemoveContainer" containerID="2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7" Oct 04 07:29:22 crc kubenswrapper[4987]: E1004 07:29:22.525944 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7\": container with ID starting with 2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7 not found: ID does not exist" containerID="2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.525993 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7"} err="failed to get container status \"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7\": rpc error: code = NotFound desc = could not find container \"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7\": container with ID starting with 2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7 not found: ID does not exist" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.526022 4987 scope.go:117] "RemoveContainer" containerID="12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5" Oct 04 07:29:22 crc kubenswrapper[4987]: E1004 07:29:22.526679 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5\": container with ID starting with 12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5 not found: ID does not exist" containerID="12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.526733 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5"} err="failed to get container status \"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5\": rpc error: code = NotFound desc = could not find container \"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5\": container with ID starting with 12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5 not found: ID does not exist" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.526771 4987 scope.go:117] "RemoveContainer" containerID="2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.527103 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7"} err="failed to get container status \"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7\": rpc error: code = NotFound desc = could not find container \"2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7\": container with ID starting with 2a77d4b988a51a123aed97d143186ee8a06cd52b080ca841140fb4469fdd48e7 not found: ID does not exist" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.527138 4987 scope.go:117] "RemoveContainer" containerID="12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.527371 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5"} err="failed to get container status \"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5\": rpc error: code = NotFound desc = could not find container \"12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5\": container with ID starting with 12d84929fd3c5bcce35af6a62272b00ef4bd8ac33e99526690c76cf983180fe5 not found: ID does not exist" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.554910 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660725 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-lib-modules\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660792 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-config-data\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660856 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-dev\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660881 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660903 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-nvme\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660928 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660949 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-httpd-run\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.660983 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-run\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.661024 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsjj2\" (UniqueName: \"kubernetes.io/projected/1546216a-0413-473d-b553-32d5c86c484a-kube-api-access-wsjj2\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.661051 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-sys\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.661071 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-scripts\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.661246 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.661271 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-logs\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.661306 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763048 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-dev\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763104 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763121 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-nvme\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763166 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763182 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-httpd-run\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763215 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-run\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763244 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsjj2\" (UniqueName: \"kubernetes.io/projected/1546216a-0413-473d-b553-32d5c86c484a-kube-api-access-wsjj2\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763248 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-dev\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763300 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-sys\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763261 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-sys\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763341 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763331 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-nvme\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763362 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-scripts\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763372 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-run\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763413 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763435 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-logs\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763491 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763535 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.763599 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.764001 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-httpd-run\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.764049 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-logs\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.764052 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.764141 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-lib-modules\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.764176 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-config-data\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.764379 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-lib-modules\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.773996 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-scripts\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.774587 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-config-data\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.786887 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.791926 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsjj2\" (UniqueName: \"kubernetes.io/projected/1546216a-0413-473d-b553-32d5c86c484a-kube-api-access-wsjj2\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.796118 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-1\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:22 crc kubenswrapper[4987]: I1004 07:29:22.872951 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:23 crc kubenswrapper[4987]: I1004 07:29:23.323702 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:29:23 crc kubenswrapper[4987]: W1004 07:29:23.334795 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1546216a_0413_473d_b553_32d5c86c484a.slice/crio-7d473be8856b6ca15e464911338599110f687d26ed09e8b2a21b9e880b687f8c WatchSource:0}: Error finding container 7d473be8856b6ca15e464911338599110f687d26ed09e8b2a21b9e880b687f8c: Status 404 returned error can't find the container with id 7d473be8856b6ca15e464911338599110f687d26ed09e8b2a21b9e880b687f8c Oct 04 07:29:23 crc kubenswrapper[4987]: I1004 07:29:23.437471 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1546216a-0413-473d-b553-32d5c86c484a","Type":"ContainerStarted","Data":"7d473be8856b6ca15e464911338599110f687d26ed09e8b2a21b9e880b687f8c"} Oct 04 07:29:24 crc kubenswrapper[4987]: I1004 07:29:24.068344 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e05172a1-4714-44c6-b500-41eed762e8ba" path="/var/lib/kubelet/pods/e05172a1-4714-44c6-b500-41eed762e8ba/volumes" Oct 04 07:29:24 crc kubenswrapper[4987]: I1004 07:29:24.452797 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1546216a-0413-473d-b553-32d5c86c484a","Type":"ContainerStarted","Data":"03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab"} Oct 04 07:29:24 crc kubenswrapper[4987]: I1004 07:29:24.452868 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1546216a-0413-473d-b553-32d5c86c484a","Type":"ContainerStarted","Data":"1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d"} Oct 04 07:29:24 crc kubenswrapper[4987]: I1004 07:29:24.481808 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.481778471 podStartE2EDuration="2.481778471s" podCreationTimestamp="2025-10-04 07:29:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:29:24.478735815 +0000 UTC m=+1181.533634667" watchObservedRunningTime="2025-10-04 07:29:24.481778471 +0000 UTC m=+1181.536677313" Oct 04 07:29:27 crc kubenswrapper[4987]: I1004 07:29:27.156470 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:27 crc kubenswrapper[4987]: I1004 07:29:27.156520 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:27 crc kubenswrapper[4987]: I1004 07:29:27.191054 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:27 crc kubenswrapper[4987]: I1004 07:29:27.221669 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:27 crc kubenswrapper[4987]: I1004 07:29:27.479479 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:27 crc kubenswrapper[4987]: I1004 07:29:27.480047 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:29 crc kubenswrapper[4987]: I1004 07:29:29.782460 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:29 crc kubenswrapper[4987]: I1004 07:29:29.784375 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:29:29 crc kubenswrapper[4987]: I1004 07:29:29.923351 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:32 crc kubenswrapper[4987]: I1004 07:29:32.873563 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:32 crc kubenswrapper[4987]: I1004 07:29:32.873704 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:32 crc kubenswrapper[4987]: I1004 07:29:32.936723 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:32 crc kubenswrapper[4987]: I1004 07:29:32.954659 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:33 crc kubenswrapper[4987]: I1004 07:29:33.534312 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:33 crc kubenswrapper[4987]: I1004 07:29:33.534815 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:35 crc kubenswrapper[4987]: I1004 07:29:35.696589 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:35 crc kubenswrapper[4987]: I1004 07:29:35.697244 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:29:35 crc kubenswrapper[4987]: I1004 07:29:35.835355 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:29:35 crc kubenswrapper[4987]: I1004 07:29:35.885701 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:35 crc kubenswrapper[4987]: I1004 07:29:35.886097 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-log" containerID="cri-o://6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb" gracePeriod=30 Oct 04 07:29:35 crc kubenswrapper[4987]: I1004 07:29:35.886443 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-httpd" containerID="cri-o://f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef" gracePeriod=30 Oct 04 07:29:36 crc kubenswrapper[4987]: I1004 07:29:36.577695 4987 generic.go:334] "Generic (PLEG): container finished" podID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerID="6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb" exitCode=143 Oct 04 07:29:36 crc kubenswrapper[4987]: I1004 07:29:36.577766 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2066e87b-2788-47e8-ba09-4a8039b07e26","Type":"ContainerDied","Data":"6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb"} Oct 04 07:29:39 crc kubenswrapper[4987]: E1004 07:29:39.136329 4987 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2066e87b_2788_47e8_ba09_4a8039b07e26.slice/crio-conmon-f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2066e87b_2788_47e8_ba09_4a8039b07e26.slice/crio-f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef.scope\": RecentStats: unable to find data in memory cache]" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.439867 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.508208 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-iscsi\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.508323 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.508932 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-var-locks-brick\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509109 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.508956 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509268 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509347 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-run\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509435 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-run" (OuterVolumeSpecName: "run") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509441 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-config-data\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509501 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-scripts\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509549 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-lib-modules\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509563 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-sys\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509577 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-nvme\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509606 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-httpd-run\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509688 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-dev\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509736 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-logs\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.509794 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ggz7\" (UniqueName: \"kubernetes.io/projected/2066e87b-2788-47e8-ba09-4a8039b07e26-kube-api-access-8ggz7\") pod \"2066e87b-2788-47e8-ba09-4a8039b07e26\" (UID: \"2066e87b-2788-47e8-ba09-4a8039b07e26\") " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510030 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510357 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-dev" (OuterVolumeSpecName: "dev") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510531 4987 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510607 4987 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510692 4987 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510769 4987 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.510837 4987 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-dev\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.511157 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.511308 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.511312 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-logs" (OuterVolumeSpecName: "logs") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.511345 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-sys" (OuterVolumeSpecName: "sys") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.516324 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.516468 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2066e87b-2788-47e8-ba09-4a8039b07e26-kube-api-access-8ggz7" (OuterVolumeSpecName: "kube-api-access-8ggz7") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "kube-api-access-8ggz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.517791 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.525226 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-scripts" (OuterVolumeSpecName: "scripts") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.558857 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-config-data" (OuterVolumeSpecName: "config-data") pod "2066e87b-2788-47e8-ba09-4a8039b07e26" (UID: "2066e87b-2788-47e8-ba09-4a8039b07e26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.603617 4987 generic.go:334] "Generic (PLEG): container finished" podID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerID="f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef" exitCode=0 Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.603733 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.603731 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2066e87b-2788-47e8-ba09-4a8039b07e26","Type":"ContainerDied","Data":"f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef"} Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.604152 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"2066e87b-2788-47e8-ba09-4a8039b07e26","Type":"ContainerDied","Data":"180f4d2e2e1627c438b01a8726061ac14028a3d78dd56612243d13a5425e7e2f"} Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.604186 4987 scope.go:117] "RemoveContainer" containerID="f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612696 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612743 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612755 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2066e87b-2788-47e8-ba09-4a8039b07e26-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612764 4987 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612774 4987 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/2066e87b-2788-47e8-ba09-4a8039b07e26-sys\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612783 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612793 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2066e87b-2788-47e8-ba09-4a8039b07e26-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612802 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ggz7\" (UniqueName: \"kubernetes.io/projected/2066e87b-2788-47e8-ba09-4a8039b07e26-kube-api-access-8ggz7\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.612822 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.631561 4987 scope.go:117] "RemoveContainer" containerID="6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.632450 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.638426 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.655080 4987 scope.go:117] "RemoveContainer" containerID="f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.657829 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:39 crc kubenswrapper[4987]: E1004 07:29:39.658320 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef\": container with ID starting with f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef not found: ID does not exist" containerID="f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.658473 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef"} err="failed to get container status \"f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef\": rpc error: code = NotFound desc = could not find container \"f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef\": container with ID starting with f2cb815f9934d225afeda667b0a29af192554d9507b2a28066c3288cf605a2ef not found: ID does not exist" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.658607 4987 scope.go:117] "RemoveContainer" containerID="6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.662043 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:39 crc kubenswrapper[4987]: E1004 07:29:39.662429 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb\": container with ID starting with 6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb not found: ID does not exist" containerID="6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.662562 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb"} err="failed to get container status \"6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb\": rpc error: code = NotFound desc = could not find container \"6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb\": container with ID starting with 6a5210a7d5915ff1a894158900b577d9485afb549ed13b08b60d4e82b21deaeb not found: ID does not exist" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.681469 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:39 crc kubenswrapper[4987]: E1004 07:29:39.687119 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-httpd" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.687166 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-httpd" Oct 04 07:29:39 crc kubenswrapper[4987]: E1004 07:29:39.687211 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-log" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.687217 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-log" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.687410 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-log" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.687424 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" containerName="glance-httpd" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.688292 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.699757 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.714674 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.714715 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817070 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-httpd-run\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817165 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817421 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-sys\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817492 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-config-data\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817520 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ns5f\" (UniqueName: \"kubernetes.io/projected/7b5a29b0-8466-4968-b2a7-a3f6bafda678-kube-api-access-9ns5f\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817618 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-run\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817747 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817816 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817890 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-logs\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.817925 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-dev\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.818077 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-scripts\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.818255 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-lib-modules\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.818313 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.818413 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919701 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-httpd-run\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919783 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919821 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-sys\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919843 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-config-data\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919875 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ns5f\" (UniqueName: \"kubernetes.io/projected/7b5a29b0-8466-4968-b2a7-a3f6bafda678-kube-api-access-9ns5f\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919911 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-run\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919960 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.919992 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920018 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-logs\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920048 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-dev\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920086 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-run\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920144 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-sys\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920101 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-scripts\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920173 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-httpd-run\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920208 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920205 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-dev\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920248 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920437 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-lib-modules\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920444 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920489 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-lib-modules\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920522 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920588 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-logs\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920615 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920656 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.920799 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.927183 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-scripts\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.927555 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-config-data\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.938127 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ns5f\" (UniqueName: \"kubernetes.io/projected/7b5a29b0-8466-4968-b2a7-a3f6bafda678-kube-api-access-9ns5f\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.942612 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:39 crc kubenswrapper[4987]: I1004 07:29:39.943613 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-single-0\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:40 crc kubenswrapper[4987]: I1004 07:29:40.006476 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:40 crc kubenswrapper[4987]: I1004 07:29:40.078360 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2066e87b-2788-47e8-ba09-4a8039b07e26" path="/var/lib/kubelet/pods/2066e87b-2788-47e8-ba09-4a8039b07e26/volumes" Oct 04 07:29:40 crc kubenswrapper[4987]: I1004 07:29:40.585235 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:29:40 crc kubenswrapper[4987]: I1004 07:29:40.614504 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7b5a29b0-8466-4968-b2a7-a3f6bafda678","Type":"ContainerStarted","Data":"0f6ea5aedb626fcea41b5301afc013f714922e98b6585d4cd83521dcb66b976a"} Oct 04 07:29:41 crc kubenswrapper[4987]: I1004 07:29:41.625046 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7b5a29b0-8466-4968-b2a7-a3f6bafda678","Type":"ContainerStarted","Data":"7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856"} Oct 04 07:29:41 crc kubenswrapper[4987]: I1004 07:29:41.625780 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7b5a29b0-8466-4968-b2a7-a3f6bafda678","Type":"ContainerStarted","Data":"81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592"} Oct 04 07:29:41 crc kubenswrapper[4987]: I1004 07:29:41.655868 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.655842226 podStartE2EDuration="2.655842226s" podCreationTimestamp="2025-10-04 07:29:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:29:41.648041211 +0000 UTC m=+1198.702940053" watchObservedRunningTime="2025-10-04 07:29:41.655842226 +0000 UTC m=+1198.710741068" Oct 04 07:29:50 crc kubenswrapper[4987]: I1004 07:29:50.007376 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:50 crc kubenswrapper[4987]: I1004 07:29:50.008190 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:50 crc kubenswrapper[4987]: I1004 07:29:50.038219 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:50 crc kubenswrapper[4987]: I1004 07:29:50.074697 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:50 crc kubenswrapper[4987]: I1004 07:29:50.703022 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:50 crc kubenswrapper[4987]: I1004 07:29:50.703078 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:52 crc kubenswrapper[4987]: I1004 07:29:52.717640 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:29:52 crc kubenswrapper[4987]: I1004 07:29:52.719493 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.226130 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt"] Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.229674 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.232850 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.233037 4987 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.243569 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj"] Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.244864 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.252698 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27"] Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.253970 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.275815 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt"] Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.295833 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj"] Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.300876 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27"] Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344423 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/660eb1ff-870f-4828-b14a-d99ab2b5659f-image-cache-config-data\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344499 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/27ab7cfc-3f81-4449-b769-9efad70c3186-image-cache-config-data\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344574 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344605 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm5jm\" (UniqueName: \"kubernetes.io/projected/660eb1ff-870f-4828-b14a-d99ab2b5659f-kube-api-access-bm5jm\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344646 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344672 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw8d8\" (UniqueName: \"kubernetes.io/projected/27ab7cfc-3f81-4449-b769-9efad70c3186-kube-api-access-lw8d8\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344695 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l9mt\" (UniqueName: \"kubernetes.io/projected/729ebc0c-e7f3-4839-9cfe-4b493b81a176-kube-api-access-7l9mt\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344715 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/729ebc0c-e7f3-4839-9cfe-4b493b81a176-config-volume\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.344751 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/729ebc0c-e7f3-4839-9cfe-4b493b81a176-secret-volume\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.389983 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.393485 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.446858 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/27ab7cfc-3f81-4449-b769-9efad70c3186-image-cache-config-data\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.447465 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5jm\" (UniqueName: \"kubernetes.io/projected/660eb1ff-870f-4828-b14a-d99ab2b5659f-kube-api-access-bm5jm\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.447645 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw8d8\" (UniqueName: \"kubernetes.io/projected/27ab7cfc-3f81-4449-b769-9efad70c3186-kube-api-access-lw8d8\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.447781 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l9mt\" (UniqueName: \"kubernetes.io/projected/729ebc0c-e7f3-4839-9cfe-4b493b81a176-kube-api-access-7l9mt\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.447898 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/729ebc0c-e7f3-4839-9cfe-4b493b81a176-config-volume\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.448037 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/729ebc0c-e7f3-4839-9cfe-4b493b81a176-secret-volume\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.448207 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/660eb1ff-870f-4828-b14a-d99ab2b5659f-image-cache-config-data\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.448777 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/729ebc0c-e7f3-4839-9cfe-4b493b81a176-config-volume\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.453483 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/660eb1ff-870f-4828-b14a-d99ab2b5659f-image-cache-config-data\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.453652 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/27ab7cfc-3f81-4449-b769-9efad70c3186-image-cache-config-data\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.456575 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/729ebc0c-e7f3-4839-9cfe-4b493b81a176-secret-volume\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.472938 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l9mt\" (UniqueName: \"kubernetes.io/projected/729ebc0c-e7f3-4839-9cfe-4b493b81a176-kube-api-access-7l9mt\") pod \"collect-profiles-29326050-422rt\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.476372 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw8d8\" (UniqueName: \"kubernetes.io/projected/27ab7cfc-3f81-4449-b769-9efad70c3186-kube-api-access-lw8d8\") pod \"glance-cache-glance-default-single-0-cleaner-29326050-sgd27\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.477871 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm5jm\" (UniqueName: \"kubernetes.io/projected/660eb1ff-870f-4828-b14a-d99ab2b5659f-kube-api-access-bm5jm\") pod \"glance-cache-glance-default-single-1-cleaner-29326050-f5gdj\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.558233 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.571723 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.586155 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:00 crc kubenswrapper[4987]: I1004 07:30:00.832330 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt"] Oct 04 07:30:00 crc kubenswrapper[4987]: W1004 07:30:00.846543 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod729ebc0c_e7f3_4839_9cfe_4b493b81a176.slice/crio-5557f79331695aaf7afee8f921f108dcb30038c5b227f6e3cf755418c06f55cd WatchSource:0}: Error finding container 5557f79331695aaf7afee8f921f108dcb30038c5b227f6e3cf755418c06f55cd: Status 404 returned error can't find the container with id 5557f79331695aaf7afee8f921f108dcb30038c5b227f6e3cf755418c06f55cd Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.081168 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj"] Oct 04 07:30:01 crc kubenswrapper[4987]: W1004 07:30:01.090954 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod660eb1ff_870f_4828_b14a_d99ab2b5659f.slice/crio-48399de028b522e80066de7c5ae46330aa5c97d844b17a55735af84eaeb8ab54 WatchSource:0}: Error finding container 48399de028b522e80066de7c5ae46330aa5c97d844b17a55735af84eaeb8ab54: Status 404 returned error can't find the container with id 48399de028b522e80066de7c5ae46330aa5c97d844b17a55735af84eaeb8ab54 Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.120266 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27"] Oct 04 07:30:01 crc kubenswrapper[4987]: W1004 07:30:01.124954 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27ab7cfc_3f81_4449_b769_9efad70c3186.slice/crio-a89e1f8bd99a04a91af62a320e9739f60178a9d0d6a9481bf69414c19eb94a15 WatchSource:0}: Error finding container a89e1f8bd99a04a91af62a320e9739f60178a9d0d6a9481bf69414c19eb94a15: Status 404 returned error can't find the container with id a89e1f8bd99a04a91af62a320e9739f60178a9d0d6a9481bf69414c19eb94a15 Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.805565 4987 generic.go:334] "Generic (PLEG): container finished" podID="729ebc0c-e7f3-4839-9cfe-4b493b81a176" containerID="5794f24591079d695193e37d7a921e848be351b5d4b6238de077eb4ed1697e79" exitCode=0 Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.805691 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" event={"ID":"729ebc0c-e7f3-4839-9cfe-4b493b81a176","Type":"ContainerDied","Data":"5794f24591079d695193e37d7a921e848be351b5d4b6238de077eb4ed1697e79"} Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.806010 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" event={"ID":"729ebc0c-e7f3-4839-9cfe-4b493b81a176","Type":"ContainerStarted","Data":"5557f79331695aaf7afee8f921f108dcb30038c5b227f6e3cf755418c06f55cd"} Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.807503 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" event={"ID":"27ab7cfc-3f81-4449-b769-9efad70c3186","Type":"ContainerStarted","Data":"64077221f913273b05f07b08432f831cece5d5d0e65bd7ebc1ccab4052a008cb"} Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.807563 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" event={"ID":"27ab7cfc-3f81-4449-b769-9efad70c3186","Type":"ContainerStarted","Data":"a89e1f8bd99a04a91af62a320e9739f60178a9d0d6a9481bf69414c19eb94a15"} Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.810150 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" event={"ID":"660eb1ff-870f-4828-b14a-d99ab2b5659f","Type":"ContainerStarted","Data":"65d310628aac2f405df3ebf65965e0164c5df4a60d34ad4c46eef77bba52ab06"} Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.810220 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" event={"ID":"660eb1ff-870f-4828-b14a-d99ab2b5659f","Type":"ContainerStarted","Data":"48399de028b522e80066de7c5ae46330aa5c97d844b17a55735af84eaeb8ab54"} Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.842116 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" podStartSLOduration=1.842091427 podStartE2EDuration="1.842091427s" podCreationTimestamp="2025-10-04 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:30:01.839914769 +0000 UTC m=+1218.894813611" watchObservedRunningTime="2025-10-04 07:30:01.842091427 +0000 UTC m=+1218.896990269" Oct 04 07:30:01 crc kubenswrapper[4987]: I1004 07:30:01.859588 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" podStartSLOduration=1.8595569969999999 podStartE2EDuration="1.859556997s" podCreationTimestamp="2025-10-04 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:30:01.85931573 +0000 UTC m=+1218.914214582" watchObservedRunningTime="2025-10-04 07:30:01.859556997 +0000 UTC m=+1218.914455839" Oct 04 07:30:02 crc kubenswrapper[4987]: I1004 07:30:02.824381 4987 generic.go:334] "Generic (PLEG): container finished" podID="27ab7cfc-3f81-4449-b769-9efad70c3186" containerID="64077221f913273b05f07b08432f831cece5d5d0e65bd7ebc1ccab4052a008cb" exitCode=0 Oct 04 07:30:02 crc kubenswrapper[4987]: I1004 07:30:02.825406 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" event={"ID":"27ab7cfc-3f81-4449-b769-9efad70c3186","Type":"ContainerDied","Data":"64077221f913273b05f07b08432f831cece5d5d0e65bd7ebc1ccab4052a008cb"} Oct 04 07:30:02 crc kubenswrapper[4987]: I1004 07:30:02.827663 4987 generic.go:334] "Generic (PLEG): container finished" podID="660eb1ff-870f-4828-b14a-d99ab2b5659f" containerID="65d310628aac2f405df3ebf65965e0164c5df4a60d34ad4c46eef77bba52ab06" exitCode=0 Oct 04 07:30:02 crc kubenswrapper[4987]: I1004 07:30:02.827775 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" event={"ID":"660eb1ff-870f-4828-b14a-d99ab2b5659f","Type":"ContainerDied","Data":"65d310628aac2f405df3ebf65965e0164c5df4a60d34ad4c46eef77bba52ab06"} Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.111301 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.217532 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l9mt\" (UniqueName: \"kubernetes.io/projected/729ebc0c-e7f3-4839-9cfe-4b493b81a176-kube-api-access-7l9mt\") pod \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.217792 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/729ebc0c-e7f3-4839-9cfe-4b493b81a176-config-volume\") pod \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.217979 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/729ebc0c-e7f3-4839-9cfe-4b493b81a176-secret-volume\") pod \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\" (UID: \"729ebc0c-e7f3-4839-9cfe-4b493b81a176\") " Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.218831 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/729ebc0c-e7f3-4839-9cfe-4b493b81a176-config-volume" (OuterVolumeSpecName: "config-volume") pod "729ebc0c-e7f3-4839-9cfe-4b493b81a176" (UID: "729ebc0c-e7f3-4839-9cfe-4b493b81a176"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.236798 4987 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/729ebc0c-e7f3-4839-9cfe-4b493b81a176-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.243019 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/729ebc0c-e7f3-4839-9cfe-4b493b81a176-kube-api-access-7l9mt" (OuterVolumeSpecName: "kube-api-access-7l9mt") pod "729ebc0c-e7f3-4839-9cfe-4b493b81a176" (UID: "729ebc0c-e7f3-4839-9cfe-4b493b81a176"). InnerVolumeSpecName "kube-api-access-7l9mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.249298 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/729ebc0c-e7f3-4839-9cfe-4b493b81a176-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "729ebc0c-e7f3-4839-9cfe-4b493b81a176" (UID: "729ebc0c-e7f3-4839-9cfe-4b493b81a176"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.338148 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l9mt\" (UniqueName: \"kubernetes.io/projected/729ebc0c-e7f3-4839-9cfe-4b493b81a176-kube-api-access-7l9mt\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.338202 4987 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/729ebc0c-e7f3-4839-9cfe-4b493b81a176-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.839286 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.839447 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326050-422rt" event={"ID":"729ebc0c-e7f3-4839-9cfe-4b493b81a176","Type":"ContainerDied","Data":"5557f79331695aaf7afee8f921f108dcb30038c5b227f6e3cf755418c06f55cd"} Oct 04 07:30:03 crc kubenswrapper[4987]: I1004 07:30:03.839879 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5557f79331695aaf7afee8f921f108dcb30038c5b227f6e3cf755418c06f55cd" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.198847 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.203560 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.255761 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/27ab7cfc-3f81-4449-b769-9efad70c3186-image-cache-config-data\") pod \"27ab7cfc-3f81-4449-b769-9efad70c3186\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.255821 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"660eb1ff-870f-4828-b14a-d99ab2b5659f\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.255857 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw8d8\" (UniqueName: \"kubernetes.io/projected/27ab7cfc-3f81-4449-b769-9efad70c3186-kube-api-access-lw8d8\") pod \"27ab7cfc-3f81-4449-b769-9efad70c3186\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.255963 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/660eb1ff-870f-4828-b14a-d99ab2b5659f-image-cache-config-data\") pod \"660eb1ff-870f-4828-b14a-d99ab2b5659f\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.255992 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"27ab7cfc-3f81-4449-b769-9efad70c3186\" (UID: \"27ab7cfc-3f81-4449-b769-9efad70c3186\") " Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.256048 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm5jm\" (UniqueName: \"kubernetes.io/projected/660eb1ff-870f-4828-b14a-d99ab2b5659f-kube-api-access-bm5jm\") pod \"660eb1ff-870f-4828-b14a-d99ab2b5659f\" (UID: \"660eb1ff-870f-4828-b14a-d99ab2b5659f\") " Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.260821 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "27ab7cfc-3f81-4449-b769-9efad70c3186" (UID: "27ab7cfc-3f81-4449-b769-9efad70c3186"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.260858 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ab7cfc-3f81-4449-b769-9efad70c3186-kube-api-access-lw8d8" (OuterVolumeSpecName: "kube-api-access-lw8d8") pod "27ab7cfc-3f81-4449-b769-9efad70c3186" (UID: "27ab7cfc-3f81-4449-b769-9efad70c3186"). InnerVolumeSpecName "kube-api-access-lw8d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.261180 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660eb1ff-870f-4828-b14a-d99ab2b5659f-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "660eb1ff-870f-4828-b14a-d99ab2b5659f" (UID: "660eb1ff-870f-4828-b14a-d99ab2b5659f"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.266807 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/660eb1ff-870f-4828-b14a-d99ab2b5659f-kube-api-access-bm5jm" (OuterVolumeSpecName: "kube-api-access-bm5jm") pod "660eb1ff-870f-4828-b14a-d99ab2b5659f" (UID: "660eb1ff-870f-4828-b14a-d99ab2b5659f"). InnerVolumeSpecName "kube-api-access-bm5jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.276606 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "660eb1ff-870f-4828-b14a-d99ab2b5659f" (UID: "660eb1ff-870f-4828-b14a-d99ab2b5659f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.276769 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27ab7cfc-3f81-4449-b769-9efad70c3186-image-cache-config-data" (OuterVolumeSpecName: "image-cache-config-data") pod "27ab7cfc-3f81-4449-b769-9efad70c3186" (UID: "27ab7cfc-3f81-4449-b769-9efad70c3186"). InnerVolumeSpecName "image-cache-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.358684 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw8d8\" (UniqueName: \"kubernetes.io/projected/27ab7cfc-3f81-4449-b769-9efad70c3186-kube-api-access-lw8d8\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.358734 4987 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/660eb1ff-870f-4828-b14a-d99ab2b5659f-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.358748 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm5jm\" (UniqueName: \"kubernetes.io/projected/660eb1ff-870f-4828-b14a-d99ab2b5659f-kube-api-access-bm5jm\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.358760 4987 reconciler_common.go:293] "Volume detached for volume \"image-cache-config-data\" (UniqueName: \"kubernetes.io/secret/27ab7cfc-3f81-4449-b769-9efad70c3186-image-cache-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.847194 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" event={"ID":"27ab7cfc-3f81-4449-b769-9efad70c3186","Type":"ContainerDied","Data":"a89e1f8bd99a04a91af62a320e9739f60178a9d0d6a9481bf69414c19eb94a15"} Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.847243 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a89e1f8bd99a04a91af62a320e9739f60178a9d0d6a9481bf69414c19eb94a15" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.847301 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.852659 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" event={"ID":"660eb1ff-870f-4828-b14a-d99ab2b5659f","Type":"ContainerDied","Data":"48399de028b522e80066de7c5ae46330aa5c97d844b17a55735af84eaeb8ab54"} Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.852689 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48399de028b522e80066de7c5ae46330aa5c97d844b17a55735af84eaeb8ab54" Oct 04 07:30:04 crc kubenswrapper[4987]: I1004 07:30:04.852739 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.670090 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nmfp4"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.676282 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nmfp4"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.758541 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancef0dc-account-delete-r5dqx"] Oct 04 07:30:17 crc kubenswrapper[4987]: E1004 07:30:17.758984 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="729ebc0c-e7f3-4839-9cfe-4b493b81a176" containerName="collect-profiles" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759014 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="729ebc0c-e7f3-4839-9cfe-4b493b81a176" containerName="collect-profiles" Oct 04 07:30:17 crc kubenswrapper[4987]: E1004 07:30:17.759085 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660eb1ff-870f-4828-b14a-d99ab2b5659f" containerName="glance-cache-glance-default-single-1-cleaner" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759098 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="660eb1ff-870f-4828-b14a-d99ab2b5659f" containerName="glance-cache-glance-default-single-1-cleaner" Oct 04 07:30:17 crc kubenswrapper[4987]: E1004 07:30:17.759134 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ab7cfc-3f81-4449-b769-9efad70c3186" containerName="glance-cache-glance-default-single-0-cleaner" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759148 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ab7cfc-3f81-4449-b769-9efad70c3186" containerName="glance-cache-glance-default-single-0-cleaner" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759308 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="729ebc0c-e7f3-4839-9cfe-4b493b81a176" containerName="collect-profiles" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759320 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ab7cfc-3f81-4449-b769-9efad70c3186" containerName="glance-cache-glance-default-single-0-cleaner" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759336 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="660eb1ff-870f-4828-b14a-d99ab2b5659f" containerName="glance-cache-glance-default-single-1-cleaner" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.759987 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.778738 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-qzmzd"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.790103 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef0dc-account-delete-r5dqx"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.795775 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-qzmzd"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.819739 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-f0dc-account-create-kskcp"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.827550 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-f0dc-account-create-kskcp"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.844712 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef0dc-account-delete-r5dqx"] Oct 04 07:30:17 crc kubenswrapper[4987]: E1004 07:30:17.845468 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-kvh4q], unattached volumes=[], failed to process volumes=[]: context canceled" pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" podUID="6c347ae4-eabf-473e-8676-8d9fd828626e" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.850343 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.850597 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="fd83077b-d8ac-4e5c-9a9d-2894d32964d6" containerName="openstackclient" containerID="cri-o://dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0" gracePeriod=30 Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.900098 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.900535 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-log" containerID="cri-o://1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d" gracePeriod=30 Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.900702 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-httpd" containerID="cri-o://03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab" gracePeriod=30 Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.901283 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvh4q\" (UniqueName: \"kubernetes.io/projected/6c347ae4-eabf-473e-8676-8d9fd828626e-kube-api-access-kvh4q\") pod \"glancef0dc-account-delete-r5dqx\" (UID: \"6c347ae4-eabf-473e-8676-8d9fd828626e\") " pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.924849 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.925156 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-log" containerID="cri-o://81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592" gracePeriod=30 Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.925655 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-httpd" containerID="cri-o://7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856" gracePeriod=30 Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.948687 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.956295 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-0-cleaner-29326050-sgd27"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.960960 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.966689 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj"] Oct 04 07:30:17 crc kubenswrapper[4987]: I1004 07:30:17.979551 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-cache-glance-default-single-1-cleaner-29326050-f5gdj"] Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.002920 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvh4q\" (UniqueName: \"kubernetes.io/projected/6c347ae4-eabf-473e-8676-8d9fd828626e-kube-api-access-kvh4q\") pod \"glancef0dc-account-delete-r5dqx\" (UID: \"6c347ae4-eabf-473e-8676-8d9fd828626e\") " pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.033639 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvh4q\" (UniqueName: \"kubernetes.io/projected/6c347ae4-eabf-473e-8676-8d9fd828626e-kube-api-access-kvh4q\") pod \"glancef0dc-account-delete-r5dqx\" (UID: \"6c347ae4-eabf-473e-8676-8d9fd828626e\") " pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.064287 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27ab7cfc-3f81-4449-b769-9efad70c3186" path="/var/lib/kubelet/pods/27ab7cfc-3f81-4449-b769-9efad70c3186/volumes" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.065146 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b53ffe9-da04-4a9f-bf00-557c359928f2" path="/var/lib/kubelet/pods/4b53ffe9-da04-4a9f-bf00-557c359928f2/volumes" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.065751 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="660eb1ff-870f-4828-b14a-d99ab2b5659f" path="/var/lib/kubelet/pods/660eb1ff-870f-4828-b14a-d99ab2b5659f/volumes" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.067736 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f" path="/var/lib/kubelet/pods/8eeb4b93-7c90-4f4e-b5bb-b2c327f3b22f/volumes" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.068612 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1343ba4-662f-4e15-a437-fb80f6987c0e" path="/var/lib/kubelet/pods/d1343ba4-662f-4e15-a437-fb80f6987c0e/volumes" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.107100 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.205653 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvh4q\" (UniqueName: \"kubernetes.io/projected/6c347ae4-eabf-473e-8676-8d9fd828626e-kube-api-access-kvh4q\") pod \"6c347ae4-eabf-473e-8676-8d9fd828626e\" (UID: \"6c347ae4-eabf-473e-8676-8d9fd828626e\") " Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.214032 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c347ae4-eabf-473e-8676-8d9fd828626e-kube-api-access-kvh4q" (OuterVolumeSpecName: "kube-api-access-kvh4q") pod "6c347ae4-eabf-473e-8676-8d9fd828626e" (UID: "6c347ae4-eabf-473e-8676-8d9fd828626e"). InnerVolumeSpecName "kube-api-access-kvh4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.308152 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvh4q\" (UniqueName: \"kubernetes.io/projected/6c347ae4-eabf-473e-8676-8d9fd828626e-kube-api-access-kvh4q\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.310047 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.408782 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nxkv\" (UniqueName: \"kubernetes.io/projected/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-kube-api-access-7nxkv\") pod \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.408847 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-scripts\") pod \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.409034 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config\") pod \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.409146 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config-secret\") pod \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\" (UID: \"fd83077b-d8ac-4e5c-9a9d-2894d32964d6\") " Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.409737 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "fd83077b-d8ac-4e5c-9a9d-2894d32964d6" (UID: "fd83077b-d8ac-4e5c-9a9d-2894d32964d6"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.409943 4987 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.414712 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-kube-api-access-7nxkv" (OuterVolumeSpecName: "kube-api-access-7nxkv") pod "fd83077b-d8ac-4e5c-9a9d-2894d32964d6" (UID: "fd83077b-d8ac-4e5c-9a9d-2894d32964d6"). InnerVolumeSpecName "kube-api-access-7nxkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.432721 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "fd83077b-d8ac-4e5c-9a9d-2894d32964d6" (UID: "fd83077b-d8ac-4e5c-9a9d-2894d32964d6"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.436548 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "fd83077b-d8ac-4e5c-9a9d-2894d32964d6" (UID: "fd83077b-d8ac-4e5c-9a9d-2894d32964d6"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.512060 4987 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.512113 4987 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.512126 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nxkv\" (UniqueName: \"kubernetes.io/projected/fd83077b-d8ac-4e5c-9a9d-2894d32964d6-kube-api-access-7nxkv\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.984594 4987 generic.go:334] "Generic (PLEG): container finished" podID="fd83077b-d8ac-4e5c-9a9d-2894d32964d6" containerID="dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0" exitCode=143 Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.984743 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"fd83077b-d8ac-4e5c-9a9d-2894d32964d6","Type":"ContainerDied","Data":"dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0"} Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.984777 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"fd83077b-d8ac-4e5c-9a9d-2894d32964d6","Type":"ContainerDied","Data":"c4763304cc3d816a1402d22b4f598663d530c3d899f792d30088361eb6344f21"} Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.984796 4987 scope.go:117] "RemoveContainer" containerID="dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.984816 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.988074 4987 generic.go:334] "Generic (PLEG): container finished" podID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerID="81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592" exitCode=143 Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.988122 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7b5a29b0-8466-4968-b2a7-a3f6bafda678","Type":"ContainerDied","Data":"81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592"} Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.990293 4987 generic.go:334] "Generic (PLEG): container finished" podID="1546216a-0413-473d-b553-32d5c86c484a" containerID="1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d" exitCode=143 Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.990325 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1546216a-0413-473d-b553-32d5c86c484a","Type":"ContainerDied","Data":"1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d"} Oct 04 07:30:18 crc kubenswrapper[4987]: I1004 07:30:18.990542 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef0dc-account-delete-r5dqx" Oct 04 07:30:19 crc kubenswrapper[4987]: I1004 07:30:19.034864 4987 scope.go:117] "RemoveContainer" containerID="dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0" Oct 04 07:30:19 crc kubenswrapper[4987]: E1004 07:30:19.035443 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0\": container with ID starting with dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0 not found: ID does not exist" containerID="dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0" Oct 04 07:30:19 crc kubenswrapper[4987]: I1004 07:30:19.035501 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0"} err="failed to get container status \"dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0\": rpc error: code = NotFound desc = could not find container \"dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0\": container with ID starting with dfbbf6bc1f224000bbbfe70b7d65d4b9a01bb970089bee95dd4e16371542c2b0 not found: ID does not exist" Oct 04 07:30:19 crc kubenswrapper[4987]: I1004 07:30:19.044461 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:30:19 crc kubenswrapper[4987]: I1004 07:30:19.059510 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:30:19 crc kubenswrapper[4987]: I1004 07:30:19.069588 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef0dc-account-delete-r5dqx"] Oct 04 07:30:19 crc kubenswrapper[4987]: I1004 07:30:19.073404 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancef0dc-account-delete-r5dqx"] Oct 04 07:30:20 crc kubenswrapper[4987]: I1004 07:30:20.063131 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c347ae4-eabf-473e-8676-8d9fd828626e" path="/var/lib/kubelet/pods/6c347ae4-eabf-473e-8676-8d9fd828626e/volumes" Oct 04 07:30:20 crc kubenswrapper[4987]: I1004 07:30:20.063961 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd83077b-d8ac-4e5c-9a9d-2894d32964d6" path="/var/lib/kubelet/pods/fd83077b-d8ac-4e5c-9a9d-2894d32964d6/volumes" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.580215 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.585938 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669340 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669407 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-dev\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669433 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-nvme\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669454 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-iscsi\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669474 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-run\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669495 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-run\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669531 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-config-data\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669553 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-dev\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669587 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-nvme\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669571 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669649 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsjj2\" (UniqueName: \"kubernetes.io/projected/1546216a-0413-473d-b553-32d5c86c484a-kube-api-access-wsjj2\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669658 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-run" (OuterVolumeSpecName: "run") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669686 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-logs\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669696 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-dev" (OuterVolumeSpecName: "dev") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669718 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-var-locks-brick\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669738 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-sys\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669726 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-run" (OuterVolumeSpecName: "run") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669767 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669799 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-config-data\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669821 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669839 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ns5f\" (UniqueName: \"kubernetes.io/projected/7b5a29b0-8466-4968-b2a7-a3f6bafda678-kube-api-access-9ns5f\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669823 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669872 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-scripts\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669917 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-logs\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669945 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-var-locks-brick\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669974 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-lib-modules\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669997 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-httpd-run\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670019 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670054 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670093 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-iscsi\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670124 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-lib-modules\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670178 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-sys\") pod \"1546216a-0413-473d-b553-32d5c86c484a\" (UID: \"1546216a-0413-473d-b553-32d5c86c484a\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670211 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-httpd-run\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670236 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-scripts\") pod \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\" (UID: \"7b5a29b0-8466-4968-b2a7-a3f6bafda678\") " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670327 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.669571 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670421 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670443 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-logs" (OuterVolumeSpecName: "logs") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670839 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-dev" (OuterVolumeSpecName: "dev") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670856 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-logs" (OuterVolumeSpecName: "logs") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670502 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670870 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670918 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-sys" (OuterVolumeSpecName: "sys") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670943 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671240 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671298 4987 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671518 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671532 4987 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671546 4987 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671560 4987 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-sys\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671576 4987 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-dev\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671618 4987 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671653 4987 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671667 4987 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671680 4987 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671695 4987 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-dev\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.670261 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-sys" (OuterVolumeSpecName: "sys") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671711 4987 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671732 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1546216a-0413-473d-b553-32d5c86c484a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671745 4987 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1546216a-0413-473d-b553-32d5c86c484a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671761 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.671774 4987 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.676840 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.676879 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b5a29b0-8466-4968-b2a7-a3f6bafda678-kube-api-access-9ns5f" (OuterVolumeSpecName: "kube-api-access-9ns5f") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "kube-api-access-9ns5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.677120 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance-cache") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.677137 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-scripts" (OuterVolumeSpecName: "scripts") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.677601 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1546216a-0413-473d-b553-32d5c86c484a-kube-api-access-wsjj2" (OuterVolumeSpecName: "kube-api-access-wsjj2") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "kube-api-access-wsjj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.678571 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-scripts" (OuterVolumeSpecName: "scripts") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.679544 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.691047 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.711166 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-config-data" (OuterVolumeSpecName: "config-data") pod "1546216a-0413-473d-b553-32d5c86c484a" (UID: "1546216a-0413-473d-b553-32d5c86c484a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.712116 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-config-data" (OuterVolumeSpecName: "config-data") pod "7b5a29b0-8466-4968-b2a7-a3f6bafda678" (UID: "7b5a29b0-8466-4968-b2a7-a3f6bafda678"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774107 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsjj2\" (UniqueName: \"kubernetes.io/projected/1546216a-0413-473d-b553-32d5c86c484a-kube-api-access-wsjj2\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774199 4987 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7b5a29b0-8466-4968-b2a7-a3f6bafda678-sys\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774258 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774303 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774323 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ns5f\" (UniqueName: \"kubernetes.io/projected/7b5a29b0-8466-4968-b2a7-a3f6bafda678-kube-api-access-9ns5f\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774342 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1546216a-0413-473d-b553-32d5c86c484a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774371 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774395 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774417 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7b5a29b0-8466-4968-b2a7-a3f6bafda678-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774434 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774459 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.774477 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b5a29b0-8466-4968-b2a7-a3f6bafda678-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.787762 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.788257 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.789325 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.802979 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.876162 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.876205 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.876214 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:21 crc kubenswrapper[4987]: I1004 07:30:21.876225 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.018845 4987 generic.go:334] "Generic (PLEG): container finished" podID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerID="7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856" exitCode=0 Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.018948 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.018951 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7b5a29b0-8466-4968-b2a7-a3f6bafda678","Type":"ContainerDied","Data":"7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856"} Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.019027 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7b5a29b0-8466-4968-b2a7-a3f6bafda678","Type":"ContainerDied","Data":"0f6ea5aedb626fcea41b5301afc013f714922e98b6585d4cd83521dcb66b976a"} Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.019121 4987 scope.go:117] "RemoveContainer" containerID="7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.022527 4987 generic.go:334] "Generic (PLEG): container finished" podID="1546216a-0413-473d-b553-32d5c86c484a" containerID="03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab" exitCode=0 Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.022673 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1546216a-0413-473d-b553-32d5c86c484a","Type":"ContainerDied","Data":"03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab"} Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.023014 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"1546216a-0413-473d-b553-32d5c86c484a","Type":"ContainerDied","Data":"7d473be8856b6ca15e464911338599110f687d26ed09e8b2a21b9e880b687f8c"} Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.022781 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.058726 4987 scope.go:117] "RemoveContainer" containerID="81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.115322 4987 scope.go:117] "RemoveContainer" containerID="7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856" Oct 04 07:30:22 crc kubenswrapper[4987]: E1004 07:30:22.123377 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856\": container with ID starting with 7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856 not found: ID does not exist" containerID="7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.124219 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856"} err="failed to get container status \"7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856\": rpc error: code = NotFound desc = could not find container \"7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856\": container with ID starting with 7ae19a231c507eaf93fbed54cc804bdfe038e049bb49abc3e183861bf0512856 not found: ID does not exist" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.124276 4987 scope.go:117] "RemoveContainer" containerID="81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592" Oct 04 07:30:22 crc kubenswrapper[4987]: E1004 07:30:22.132159 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592\": container with ID starting with 81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592 not found: ID does not exist" containerID="81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.132216 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592"} err="failed to get container status \"81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592\": rpc error: code = NotFound desc = could not find container \"81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592\": container with ID starting with 81f3886ee1c8160f92cb78916a501d42704378900d96e9bcb135b5001ff06592 not found: ID does not exist" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.132363 4987 scope.go:117] "RemoveContainer" containerID="03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.152736 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.187149 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.220130 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.223843 4987 scope.go:117] "RemoveContainer" containerID="1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.232561 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.240870 4987 scope.go:117] "RemoveContainer" containerID="03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab" Oct 04 07:30:22 crc kubenswrapper[4987]: E1004 07:30:22.244268 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab\": container with ID starting with 03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab not found: ID does not exist" containerID="03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.244323 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab"} err="failed to get container status \"03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab\": rpc error: code = NotFound desc = could not find container \"03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab\": container with ID starting with 03a40d650100cbfd62fa934f89043bbfdf35e1541330646d3bd51c6521e3e9ab not found: ID does not exist" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.244354 4987 scope.go:117] "RemoveContainer" containerID="1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d" Oct 04 07:30:22 crc kubenswrapper[4987]: E1004 07:30:22.244841 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d\": container with ID starting with 1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d not found: ID does not exist" containerID="1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d" Oct 04 07:30:22 crc kubenswrapper[4987]: I1004 07:30:22.244881 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d"} err="failed to get container status \"1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d\": rpc error: code = NotFound desc = could not find container \"1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d\": container with ID starting with 1a443e4d2a155959b3dfa1bd9491c16d74ec1634e6269b3ad53c11ddea461c9d not found: ID does not exist" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.876568 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-kjbtd"] Oct 04 07:30:23 crc kubenswrapper[4987]: E1004 07:30:23.876965 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-log" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.876981 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-log" Oct 04 07:30:23 crc kubenswrapper[4987]: E1004 07:30:23.877023 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-httpd" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877032 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-httpd" Oct 04 07:30:23 crc kubenswrapper[4987]: E1004 07:30:23.877055 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-log" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877062 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-log" Oct 04 07:30:23 crc kubenswrapper[4987]: E1004 07:30:23.877095 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd83077b-d8ac-4e5c-9a9d-2894d32964d6" containerName="openstackclient" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877103 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd83077b-d8ac-4e5c-9a9d-2894d32964d6" containerName="openstackclient" Oct 04 07:30:23 crc kubenswrapper[4987]: E1004 07:30:23.877132 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-httpd" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877139 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-httpd" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877296 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-log" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877308 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd83077b-d8ac-4e5c-9a9d-2894d32964d6" containerName="openstackclient" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877318 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-log" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877327 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="1546216a-0413-473d-b553-32d5c86c484a" containerName="glance-httpd" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877348 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" containerName="glance-httpd" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.877952 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.892164 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-kjbtd"] Oct 04 07:30:23 crc kubenswrapper[4987]: I1004 07:30:23.929222 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv7pj\" (UniqueName: \"kubernetes.io/projected/3d181af4-dc7d-44e3-bb66-a06f0079ab7f-kube-api-access-zv7pj\") pod \"glance-db-create-kjbtd\" (UID: \"3d181af4-dc7d-44e3-bb66-a06f0079ab7f\") " pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:24 crc kubenswrapper[4987]: I1004 07:30:24.030956 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv7pj\" (UniqueName: \"kubernetes.io/projected/3d181af4-dc7d-44e3-bb66-a06f0079ab7f-kube-api-access-zv7pj\") pod \"glance-db-create-kjbtd\" (UID: \"3d181af4-dc7d-44e3-bb66-a06f0079ab7f\") " pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:24 crc kubenswrapper[4987]: I1004 07:30:24.061749 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv7pj\" (UniqueName: \"kubernetes.io/projected/3d181af4-dc7d-44e3-bb66-a06f0079ab7f-kube-api-access-zv7pj\") pod \"glance-db-create-kjbtd\" (UID: \"3d181af4-dc7d-44e3-bb66-a06f0079ab7f\") " pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:24 crc kubenswrapper[4987]: I1004 07:30:24.061878 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1546216a-0413-473d-b553-32d5c86c484a" path="/var/lib/kubelet/pods/1546216a-0413-473d-b553-32d5c86c484a/volumes" Oct 04 07:30:24 crc kubenswrapper[4987]: I1004 07:30:24.062495 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b5a29b0-8466-4968-b2a7-a3f6bafda678" path="/var/lib/kubelet/pods/7b5a29b0-8466-4968-b2a7-a3f6bafda678/volumes" Oct 04 07:30:24 crc kubenswrapper[4987]: I1004 07:30:24.193923 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:24 crc kubenswrapper[4987]: I1004 07:30:24.608818 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-kjbtd"] Oct 04 07:30:25 crc kubenswrapper[4987]: I1004 07:30:25.047417 4987 generic.go:334] "Generic (PLEG): container finished" podID="3d181af4-dc7d-44e3-bb66-a06f0079ab7f" containerID="230f7148684aca1d5c37104a266592388ff589f87dcb05b50527544b44898f6f" exitCode=0 Oct 04 07:30:25 crc kubenswrapper[4987]: I1004 07:30:25.047492 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-kjbtd" event={"ID":"3d181af4-dc7d-44e3-bb66-a06f0079ab7f","Type":"ContainerDied","Data":"230f7148684aca1d5c37104a266592388ff589f87dcb05b50527544b44898f6f"} Oct 04 07:30:25 crc kubenswrapper[4987]: I1004 07:30:25.047551 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-kjbtd" event={"ID":"3d181af4-dc7d-44e3-bb66-a06f0079ab7f","Type":"ContainerStarted","Data":"0ccb572c3280a4fe001d36f3b3c32f56b61a54b4d3167be40ebb7981fd3cdf15"} Oct 04 07:30:26 crc kubenswrapper[4987]: I1004 07:30:26.420957 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:26 crc kubenswrapper[4987]: I1004 07:30:26.481477 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv7pj\" (UniqueName: \"kubernetes.io/projected/3d181af4-dc7d-44e3-bb66-a06f0079ab7f-kube-api-access-zv7pj\") pod \"3d181af4-dc7d-44e3-bb66-a06f0079ab7f\" (UID: \"3d181af4-dc7d-44e3-bb66-a06f0079ab7f\") " Oct 04 07:30:26 crc kubenswrapper[4987]: I1004 07:30:26.497954 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d181af4-dc7d-44e3-bb66-a06f0079ab7f-kube-api-access-zv7pj" (OuterVolumeSpecName: "kube-api-access-zv7pj") pod "3d181af4-dc7d-44e3-bb66-a06f0079ab7f" (UID: "3d181af4-dc7d-44e3-bb66-a06f0079ab7f"). InnerVolumeSpecName "kube-api-access-zv7pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:26 crc kubenswrapper[4987]: I1004 07:30:26.584001 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv7pj\" (UniqueName: \"kubernetes.io/projected/3d181af4-dc7d-44e3-bb66-a06f0079ab7f-kube-api-access-zv7pj\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:27 crc kubenswrapper[4987]: I1004 07:30:27.062529 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-kjbtd" event={"ID":"3d181af4-dc7d-44e3-bb66-a06f0079ab7f","Type":"ContainerDied","Data":"0ccb572c3280a4fe001d36f3b3c32f56b61a54b4d3167be40ebb7981fd3cdf15"} Oct 04 07:30:27 crc kubenswrapper[4987]: I1004 07:30:27.062592 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ccb572c3280a4fe001d36f3b3c32f56b61a54b4d3167be40ebb7981fd3cdf15" Oct 04 07:30:27 crc kubenswrapper[4987]: I1004 07:30:27.062604 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-kjbtd" Oct 04 07:30:33 crc kubenswrapper[4987]: I1004 07:30:33.905734 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-9d75-account-create-fqxfz"] Oct 04 07:30:33 crc kubenswrapper[4987]: E1004 07:30:33.906866 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d181af4-dc7d-44e3-bb66-a06f0079ab7f" containerName="mariadb-database-create" Oct 04 07:30:33 crc kubenswrapper[4987]: I1004 07:30:33.906887 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d181af4-dc7d-44e3-bb66-a06f0079ab7f" containerName="mariadb-database-create" Oct 04 07:30:33 crc kubenswrapper[4987]: I1004 07:30:33.907086 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d181af4-dc7d-44e3-bb66-a06f0079ab7f" containerName="mariadb-database-create" Oct 04 07:30:33 crc kubenswrapper[4987]: I1004 07:30:33.907638 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:33 crc kubenswrapper[4987]: I1004 07:30:33.910515 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 04 07:30:33 crc kubenswrapper[4987]: I1004 07:30:33.916321 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9d75-account-create-fqxfz"] Oct 04 07:30:34 crc kubenswrapper[4987]: I1004 07:30:34.003139 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9srh6\" (UniqueName: \"kubernetes.io/projected/01ed86d9-572f-4cc3-9a45-a299e9ce89d0-kube-api-access-9srh6\") pod \"glance-9d75-account-create-fqxfz\" (UID: \"01ed86d9-572f-4cc3-9a45-a299e9ce89d0\") " pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:34 crc kubenswrapper[4987]: I1004 07:30:34.104889 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9srh6\" (UniqueName: \"kubernetes.io/projected/01ed86d9-572f-4cc3-9a45-a299e9ce89d0-kube-api-access-9srh6\") pod \"glance-9d75-account-create-fqxfz\" (UID: \"01ed86d9-572f-4cc3-9a45-a299e9ce89d0\") " pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:34 crc kubenswrapper[4987]: I1004 07:30:34.127803 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9srh6\" (UniqueName: \"kubernetes.io/projected/01ed86d9-572f-4cc3-9a45-a299e9ce89d0-kube-api-access-9srh6\") pod \"glance-9d75-account-create-fqxfz\" (UID: \"01ed86d9-572f-4cc3-9a45-a299e9ce89d0\") " pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:34 crc kubenswrapper[4987]: I1004 07:30:34.271117 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:36 crc kubenswrapper[4987]: I1004 07:30:36.339242 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9d75-account-create-fqxfz"] Oct 04 07:30:37 crc kubenswrapper[4987]: I1004 07:30:37.146138 4987 generic.go:334] "Generic (PLEG): container finished" podID="01ed86d9-572f-4cc3-9a45-a299e9ce89d0" containerID="04fdc735ad6ebffb6c35ed6aa6c6333b4e7f90a581a8fe4baec5f6aeeb3a50c3" exitCode=0 Oct 04 07:30:37 crc kubenswrapper[4987]: I1004 07:30:37.146265 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" event={"ID":"01ed86d9-572f-4cc3-9a45-a299e9ce89d0","Type":"ContainerDied","Data":"04fdc735ad6ebffb6c35ed6aa6c6333b4e7f90a581a8fe4baec5f6aeeb3a50c3"} Oct 04 07:30:37 crc kubenswrapper[4987]: I1004 07:30:37.146400 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" event={"ID":"01ed86d9-572f-4cc3-9a45-a299e9ce89d0","Type":"ContainerStarted","Data":"10bf687dae8f1a2767364fb8768ec47af41442b23351b302cf61b2c17bae536a"} Oct 04 07:30:38 crc kubenswrapper[4987]: I1004 07:30:38.520226 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:38 crc kubenswrapper[4987]: I1004 07:30:38.577173 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9srh6\" (UniqueName: \"kubernetes.io/projected/01ed86d9-572f-4cc3-9a45-a299e9ce89d0-kube-api-access-9srh6\") pod \"01ed86d9-572f-4cc3-9a45-a299e9ce89d0\" (UID: \"01ed86d9-572f-4cc3-9a45-a299e9ce89d0\") " Oct 04 07:30:38 crc kubenswrapper[4987]: I1004 07:30:38.597642 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ed86d9-572f-4cc3-9a45-a299e9ce89d0-kube-api-access-9srh6" (OuterVolumeSpecName: "kube-api-access-9srh6") pod "01ed86d9-572f-4cc3-9a45-a299e9ce89d0" (UID: "01ed86d9-572f-4cc3-9a45-a299e9ce89d0"). InnerVolumeSpecName "kube-api-access-9srh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:38 crc kubenswrapper[4987]: I1004 07:30:38.679991 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9srh6\" (UniqueName: \"kubernetes.io/projected/01ed86d9-572f-4cc3-9a45-a299e9ce89d0-kube-api-access-9srh6\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:39 crc kubenswrapper[4987]: I1004 07:30:39.166085 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" event={"ID":"01ed86d9-572f-4cc3-9a45-a299e9ce89d0","Type":"ContainerDied","Data":"10bf687dae8f1a2767364fb8768ec47af41442b23351b302cf61b2c17bae536a"} Oct 04 07:30:39 crc kubenswrapper[4987]: I1004 07:30:39.166146 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10bf687dae8f1a2767364fb8768ec47af41442b23351b302cf61b2c17bae536a" Oct 04 07:30:39 crc kubenswrapper[4987]: I1004 07:30:39.166175 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9d75-account-create-fqxfz" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.957909 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-hgnp8"] Oct 04 07:30:43 crc kubenswrapper[4987]: E1004 07:30:43.960058 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ed86d9-572f-4cc3-9a45-a299e9ce89d0" containerName="mariadb-account-create" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.960128 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ed86d9-572f-4cc3-9a45-a299e9ce89d0" containerName="mariadb-account-create" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.960364 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ed86d9-572f-4cc3-9a45-a299e9ce89d0" containerName="mariadb-account-create" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.961108 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.966310 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.966326 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.971869 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qf4m7" Oct 04 07:30:43 crc kubenswrapper[4987]: I1004 07:30:43.972801 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hgnp8"] Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.086613 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pgrv\" (UniqueName: \"kubernetes.io/projected/517a3399-3cf8-4b6d-a4bc-483b67088ba1-kube-api-access-8pgrv\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.086697 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-config-data\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.086730 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-db-sync-config-data\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.086821 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-combined-ca-bundle\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.188401 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pgrv\" (UniqueName: \"kubernetes.io/projected/517a3399-3cf8-4b6d-a4bc-483b67088ba1-kube-api-access-8pgrv\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.188495 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-config-data\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.188549 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-db-sync-config-data\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.188608 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-combined-ca-bundle\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.191913 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.192057 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.205766 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-config-data\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.207032 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-combined-ca-bundle\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.212496 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pgrv\" (UniqueName: \"kubernetes.io/projected/517a3399-3cf8-4b6d-a4bc-483b67088ba1-kube-api-access-8pgrv\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.229902 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-db-sync-config-data\") pod \"glance-db-sync-hgnp8\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.287917 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qf4m7" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.296161 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:44 crc kubenswrapper[4987]: I1004 07:30:44.737364 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hgnp8"] Oct 04 07:30:45 crc kubenswrapper[4987]: I1004 07:30:45.210426 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hgnp8" event={"ID":"517a3399-3cf8-4b6d-a4bc-483b67088ba1","Type":"ContainerStarted","Data":"7a52302e289bf3e69f14e43bad65f5fb1b135ea9979d233c31282f838bb47819"} Oct 04 07:30:46 crc kubenswrapper[4987]: I1004 07:30:46.219665 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hgnp8" event={"ID":"517a3399-3cf8-4b6d-a4bc-483b67088ba1","Type":"ContainerStarted","Data":"420ec6669ee583256e6cd9c5022954413875625e228f48012862a95cbadd77ff"} Oct 04 07:30:46 crc kubenswrapper[4987]: I1004 07:30:46.240740 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-hgnp8" podStartSLOduration=3.240712087 podStartE2EDuration="3.240712087s" podCreationTimestamp="2025-10-04 07:30:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:30:46.235919496 +0000 UTC m=+1263.290818348" watchObservedRunningTime="2025-10-04 07:30:46.240712087 +0000 UTC m=+1263.295610929" Oct 04 07:30:49 crc kubenswrapper[4987]: I1004 07:30:49.249462 4987 generic.go:334] "Generic (PLEG): container finished" podID="517a3399-3cf8-4b6d-a4bc-483b67088ba1" containerID="420ec6669ee583256e6cd9c5022954413875625e228f48012862a95cbadd77ff" exitCode=0 Oct 04 07:30:49 crc kubenswrapper[4987]: I1004 07:30:49.249715 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hgnp8" event={"ID":"517a3399-3cf8-4b6d-a4bc-483b67088ba1","Type":"ContainerDied","Data":"420ec6669ee583256e6cd9c5022954413875625e228f48012862a95cbadd77ff"} Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.588330 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.706855 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pgrv\" (UniqueName: \"kubernetes.io/projected/517a3399-3cf8-4b6d-a4bc-483b67088ba1-kube-api-access-8pgrv\") pod \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.707834 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-db-sync-config-data\") pod \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.708069 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-config-data\") pod \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.708306 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-combined-ca-bundle\") pod \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\" (UID: \"517a3399-3cf8-4b6d-a4bc-483b67088ba1\") " Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.714086 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "517a3399-3cf8-4b6d-a4bc-483b67088ba1" (UID: "517a3399-3cf8-4b6d-a4bc-483b67088ba1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.714634 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/517a3399-3cf8-4b6d-a4bc-483b67088ba1-kube-api-access-8pgrv" (OuterVolumeSpecName: "kube-api-access-8pgrv") pod "517a3399-3cf8-4b6d-a4bc-483b67088ba1" (UID: "517a3399-3cf8-4b6d-a4bc-483b67088ba1"). InnerVolumeSpecName "kube-api-access-8pgrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.741721 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "517a3399-3cf8-4b6d-a4bc-483b67088ba1" (UID: "517a3399-3cf8-4b6d-a4bc-483b67088ba1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.748514 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-config-data" (OuterVolumeSpecName: "config-data") pod "517a3399-3cf8-4b6d-a4bc-483b67088ba1" (UID: "517a3399-3cf8-4b6d-a4bc-483b67088ba1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.811038 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pgrv\" (UniqueName: \"kubernetes.io/projected/517a3399-3cf8-4b6d-a4bc-483b67088ba1-kube-api-access-8pgrv\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.811079 4987 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.811092 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:50 crc kubenswrapper[4987]: I1004 07:30:50.811109 4987 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/517a3399-3cf8-4b6d-a4bc-483b67088ba1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.266842 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hgnp8" event={"ID":"517a3399-3cf8-4b6d-a4bc-483b67088ba1","Type":"ContainerDied","Data":"7a52302e289bf3e69f14e43bad65f5fb1b135ea9979d233c31282f838bb47819"} Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.266894 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a52302e289bf3e69f14e43bad65f5fb1b135ea9979d233c31282f838bb47819" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.267343 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hgnp8" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.591675 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:51 crc kubenswrapper[4987]: E1004 07:30:51.592097 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="517a3399-3cf8-4b6d-a4bc-483b67088ba1" containerName="glance-db-sync" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.592112 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="517a3399-3cf8-4b6d-a4bc-483b67088ba1" containerName="glance-db-sync" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.592247 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="517a3399-3cf8-4b6d-a4bc-483b67088ba1" containerName="glance-db-sync" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.593055 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.595528 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.597006 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.597099 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.597006 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qf4m7" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.597159 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.597462 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.615872 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625457 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-httpd-run\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625521 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-scripts\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625543 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625662 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-logs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625717 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7bkq\" (UniqueName: \"kubernetes.io/projected/79387c21-da76-4569-8a74-31fc5eebe548-kube-api-access-p7bkq\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625824 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-config-data\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.625856 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.626007 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.626094 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727437 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7bkq\" (UniqueName: \"kubernetes.io/projected/79387c21-da76-4569-8a74-31fc5eebe548-kube-api-access-p7bkq\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727740 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-config-data\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727771 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727802 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727835 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727868 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-httpd-run\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727892 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-scripts\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727914 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.727962 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-logs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.728552 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-logs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.728929 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-httpd-run\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.729490 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.733598 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-scripts\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.733607 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.734372 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.741022 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.741474 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-config-data\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.746994 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7bkq\" (UniqueName: \"kubernetes.io/projected/79387c21-da76-4569-8a74-31fc5eebe548-kube-api-access-p7bkq\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.763437 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:51 crc kubenswrapper[4987]: I1004 07:30:51.910761 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:52 crc kubenswrapper[4987]: I1004 07:30:52.319003 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:53 crc kubenswrapper[4987]: I1004 07:30:53.163777 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:53 crc kubenswrapper[4987]: I1004 07:30:53.283970 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"79387c21-da76-4569-8a74-31fc5eebe548","Type":"ContainerStarted","Data":"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b"} Oct 04 07:30:53 crc kubenswrapper[4987]: I1004 07:30:53.284601 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"79387c21-da76-4569-8a74-31fc5eebe548","Type":"ContainerStarted","Data":"ca8230ac159f11c651cbc3f7581fadce1c602d0faaad7e79e5f964b77a5d4447"} Oct 04 07:30:54 crc kubenswrapper[4987]: I1004 07:30:54.293175 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"79387c21-da76-4569-8a74-31fc5eebe548","Type":"ContainerStarted","Data":"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea"} Oct 04 07:30:54 crc kubenswrapper[4987]: I1004 07:30:54.293367 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-log" containerID="cri-o://8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b" gracePeriod=30 Oct 04 07:30:54 crc kubenswrapper[4987]: I1004 07:30:54.293454 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-httpd" containerID="cri-o://fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea" gracePeriod=30 Oct 04 07:30:54 crc kubenswrapper[4987]: I1004 07:30:54.315742 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.315720336 podStartE2EDuration="3.315720336s" podCreationTimestamp="2025-10-04 07:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:30:54.31367523 +0000 UTC m=+1271.368574092" watchObservedRunningTime="2025-10-04 07:30:54.315720336 +0000 UTC m=+1271.370619188" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.276006 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308300 4987 generic.go:334] "Generic (PLEG): container finished" podID="79387c21-da76-4569-8a74-31fc5eebe548" containerID="fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea" exitCode=0 Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308343 4987 generic.go:334] "Generic (PLEG): container finished" podID="79387c21-da76-4569-8a74-31fc5eebe548" containerID="8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b" exitCode=143 Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308367 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"79387c21-da76-4569-8a74-31fc5eebe548","Type":"ContainerDied","Data":"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea"} Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308400 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"79387c21-da76-4569-8a74-31fc5eebe548","Type":"ContainerDied","Data":"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b"} Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308410 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"79387c21-da76-4569-8a74-31fc5eebe548","Type":"ContainerDied","Data":"ca8230ac159f11c651cbc3f7581fadce1c602d0faaad7e79e5f964b77a5d4447"} Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308427 4987 scope.go:117] "RemoveContainer" containerID="fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.308555 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.345224 4987 scope.go:117] "RemoveContainer" containerID="8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.364293 4987 scope.go:117] "RemoveContainer" containerID="fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea" Oct 04 07:30:55 crc kubenswrapper[4987]: E1004 07:30:55.364853 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea\": container with ID starting with fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea not found: ID does not exist" containerID="fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.364900 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea"} err="failed to get container status \"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea\": rpc error: code = NotFound desc = could not find container \"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea\": container with ID starting with fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea not found: ID does not exist" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.364935 4987 scope.go:117] "RemoveContainer" containerID="8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b" Oct 04 07:30:55 crc kubenswrapper[4987]: E1004 07:30:55.365413 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b\": container with ID starting with 8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b not found: ID does not exist" containerID="8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.365491 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b"} err="failed to get container status \"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b\": rpc error: code = NotFound desc = could not find container \"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b\": container with ID starting with 8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b not found: ID does not exist" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.365557 4987 scope.go:117] "RemoveContainer" containerID="fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.366206 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea"} err="failed to get container status \"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea\": rpc error: code = NotFound desc = could not find container \"fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea\": container with ID starting with fa8de4203fade3362b9ee81fd4c4dcd37042a34d098720580c5aab21021d46ea not found: ID does not exist" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.366233 4987 scope.go:117] "RemoveContainer" containerID="8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.366522 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b"} err="failed to get container status \"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b\": rpc error: code = NotFound desc = could not find container \"8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b\": container with ID starting with 8e8245131d7c418e145275ca7ce739cca97651529f4c9de6890becde03cc152b not found: ID does not exist" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383123 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-config-data\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383207 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-internal-tls-certs\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383231 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-combined-ca-bundle\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383249 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383270 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-scripts\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383328 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7bkq\" (UniqueName: \"kubernetes.io/projected/79387c21-da76-4569-8a74-31fc5eebe548-kube-api-access-p7bkq\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383371 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-httpd-run\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383404 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-public-tls-certs\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.383434 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-logs\") pod \"79387c21-da76-4569-8a74-31fc5eebe548\" (UID: \"79387c21-da76-4569-8a74-31fc5eebe548\") " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.384106 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-logs" (OuterVolumeSpecName: "logs") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.384348 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.390075 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-scripts" (OuterVolumeSpecName: "scripts") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.399864 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79387c21-da76-4569-8a74-31fc5eebe548-kube-api-access-p7bkq" (OuterVolumeSpecName: "kube-api-access-p7bkq") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "kube-api-access-p7bkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.401810 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.417654 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.422881 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.424928 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-config-data" (OuterVolumeSpecName: "config-data") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.429767 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "79387c21-da76-4569-8a74-31fc5eebe548" (UID: "79387c21-da76-4569-8a74-31fc5eebe548"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485382 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485421 4987 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485435 4987 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485478 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485487 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485497 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7bkq\" (UniqueName: \"kubernetes.io/projected/79387c21-da76-4569-8a74-31fc5eebe548-kube-api-access-p7bkq\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485509 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485518 4987 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79387c21-da76-4569-8a74-31fc5eebe548-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.485528 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/79387c21-da76-4569-8a74-31fc5eebe548-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.498702 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.586815 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.653993 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.663127 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.673870 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:55 crc kubenswrapper[4987]: E1004 07:30:55.674298 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-log" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.674318 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-log" Oct 04 07:30:55 crc kubenswrapper[4987]: E1004 07:30:55.674334 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-httpd" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.674343 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-httpd" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.674545 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-httpd" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.674575 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="79387c21-da76-4569-8a74-31fc5eebe548" containerName="glance-log" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.675508 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.677597 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.678459 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.680716 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qf4m7" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.681119 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.681661 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.681850 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.692399 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790669 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-httpd-run\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790755 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790858 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790896 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-logs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790916 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790952 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.790976 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.791130 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.791190 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k69g\" (UniqueName: \"kubernetes.io/projected/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-kube-api-access-2k69g\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893050 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893107 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-logs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893126 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893153 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893170 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893198 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893215 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k69g\" (UniqueName: \"kubernetes.io/projected/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-kube-api-access-2k69g\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893235 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-httpd-run\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893256 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893900 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-logs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.893958 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.894259 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-httpd-run\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.898965 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.899360 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.900025 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.900266 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.900488 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.917664 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k69g\" (UniqueName: \"kubernetes.io/projected/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-kube-api-access-2k69g\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.917908 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-single-0\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:55 crc kubenswrapper[4987]: I1004 07:30:55.997166 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:30:56 crc kubenswrapper[4987]: I1004 07:30:56.070163 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79387c21-da76-4569-8a74-31fc5eebe548" path="/var/lib/kubelet/pods/79387c21-da76-4569-8a74-31fc5eebe548/volumes" Oct 04 07:30:56 crc kubenswrapper[4987]: I1004 07:30:56.238979 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:30:56 crc kubenswrapper[4987]: W1004 07:30:56.243849 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod462acbe8_3917_4ac3_afab_ca02a9ee0cb7.slice/crio-43a558703ac04027ede4d2c8e502ddc0a2617137bda077d16626a96401d40fd3 WatchSource:0}: Error finding container 43a558703ac04027ede4d2c8e502ddc0a2617137bda077d16626a96401d40fd3: Status 404 returned error can't find the container with id 43a558703ac04027ede4d2c8e502ddc0a2617137bda077d16626a96401d40fd3 Oct 04 07:30:56 crc kubenswrapper[4987]: I1004 07:30:56.327489 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"462acbe8-3917-4ac3-afab-ca02a9ee0cb7","Type":"ContainerStarted","Data":"43a558703ac04027ede4d2c8e502ddc0a2617137bda077d16626a96401d40fd3"} Oct 04 07:30:57 crc kubenswrapper[4987]: I1004 07:30:57.338312 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"462acbe8-3917-4ac3-afab-ca02a9ee0cb7","Type":"ContainerStarted","Data":"b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f"} Oct 04 07:30:58 crc kubenswrapper[4987]: I1004 07:30:58.351209 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"462acbe8-3917-4ac3-afab-ca02a9ee0cb7","Type":"ContainerStarted","Data":"37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234"} Oct 04 07:30:58 crc kubenswrapper[4987]: I1004 07:30:58.384605 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.384580544 podStartE2EDuration="3.384580544s" podCreationTimestamp="2025-10-04 07:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:30:58.37720891 +0000 UTC m=+1275.432107752" watchObservedRunningTime="2025-10-04 07:30:58.384580544 +0000 UTC m=+1275.439479376" Oct 04 07:31:05 crc kubenswrapper[4987]: I1004 07:31:05.998081 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:05 crc kubenswrapper[4987]: I1004 07:31:05.998416 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:06 crc kubenswrapper[4987]: I1004 07:31:06.028059 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:06 crc kubenswrapper[4987]: I1004 07:31:06.040087 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:06 crc kubenswrapper[4987]: I1004 07:31:06.428107 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:06 crc kubenswrapper[4987]: I1004 07:31:06.428193 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:08 crc kubenswrapper[4987]: I1004 07:31:08.471246 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:31:08 crc kubenswrapper[4987]: I1004 07:31:08.471699 4987 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 07:31:08 crc kubenswrapper[4987]: I1004 07:31:08.621457 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:08 crc kubenswrapper[4987]: I1004 07:31:08.623994 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.087152 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hgnp8"] Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.099974 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hgnp8"] Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.133838 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance9d75-account-delete-559xm"] Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.134924 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.195345 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.225099 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance9d75-account-delete-559xm"] Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.243104 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bt68\" (UniqueName: \"kubernetes.io/projected/3bcaa891-328e-43c5-9ee7-748a5097cb0f-kube-api-access-4bt68\") pod \"glance9d75-account-delete-559xm\" (UID: \"3bcaa891-328e-43c5-9ee7-748a5097cb0f\") " pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.344688 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bt68\" (UniqueName: \"kubernetes.io/projected/3bcaa891-328e-43c5-9ee7-748a5097cb0f-kube-api-access-4bt68\") pod \"glance9d75-account-delete-559xm\" (UID: \"3bcaa891-328e-43c5-9ee7-748a5097cb0f\") " pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.370876 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bt68\" (UniqueName: \"kubernetes.io/projected/3bcaa891-328e-43c5-9ee7-748a5097cb0f-kube-api-access-4bt68\") pod \"glance9d75-account-delete-559xm\" (UID: \"3bcaa891-328e-43c5-9ee7-748a5097cb0f\") " pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.464602 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.493805 4987 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/glance-default-single-0" secret="" err="secret \"glance-glance-dockercfg-qf4m7\" not found" Oct 04 07:31:10 crc kubenswrapper[4987]: E1004 07:31:10.547819 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 07:31:10 crc kubenswrapper[4987]: E1004 07:31:10.547938 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 07:31:10 crc kubenswrapper[4987]: E1004 07:31:10.548341 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:11.048313641 +0000 UTC m=+1288.103212473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-scripts" not found Oct 04 07:31:10 crc kubenswrapper[4987]: E1004 07:31:10.548437 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:11.048413384 +0000 UTC m=+1288.103312226 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-default-single-config-data" not found Oct 04 07:31:10 crc kubenswrapper[4987]: I1004 07:31:10.949559 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance9d75-account-delete-559xm"] Oct 04 07:31:11 crc kubenswrapper[4987]: E1004 07:31:11.058646 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 07:31:11 crc kubenswrapper[4987]: E1004 07:31:11.058745 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:12.058725477 +0000 UTC m=+1289.113624319 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-default-single-config-data" not found Oct 04 07:31:11 crc kubenswrapper[4987]: E1004 07:31:11.058651 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 07:31:11 crc kubenswrapper[4987]: E1004 07:31:11.058799 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:12.058787848 +0000 UTC m=+1289.113686690 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-scripts" not found Oct 04 07:31:11 crc kubenswrapper[4987]: I1004 07:31:11.504923 4987 generic.go:334] "Generic (PLEG): container finished" podID="3bcaa891-328e-43c5-9ee7-748a5097cb0f" containerID="6ae61230aa2866af4eb23996387d0c3b70dc594f1a9694ccb5009cfc3fe0abf8" exitCode=0 Oct 04 07:31:11 crc kubenswrapper[4987]: I1004 07:31:11.504996 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" event={"ID":"3bcaa891-328e-43c5-9ee7-748a5097cb0f","Type":"ContainerDied","Data":"6ae61230aa2866af4eb23996387d0c3b70dc594f1a9694ccb5009cfc3fe0abf8"} Oct 04 07:31:11 crc kubenswrapper[4987]: I1004 07:31:11.505060 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" event={"ID":"3bcaa891-328e-43c5-9ee7-748a5097cb0f","Type":"ContainerStarted","Data":"958a0ec940ec5f7bd8859c0a41ffbbeaf91b27b631ca8523188644f673372d5b"} Oct 04 07:31:11 crc kubenswrapper[4987]: I1004 07:31:11.505240 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-log" containerID="cri-o://b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f" gracePeriod=30 Oct 04 07:31:11 crc kubenswrapper[4987]: I1004 07:31:11.505287 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-httpd" containerID="cri-o://37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234" gracePeriod=30 Oct 04 07:31:12 crc kubenswrapper[4987]: I1004 07:31:12.063323 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="517a3399-3cf8-4b6d-a4bc-483b67088ba1" path="/var/lib/kubelet/pods/517a3399-3cf8-4b6d-a4bc-483b67088ba1/volumes" Oct 04 07:31:12 crc kubenswrapper[4987]: E1004 07:31:12.074431 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 07:31:12 crc kubenswrapper[4987]: E1004 07:31:12.074516 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:14.074496418 +0000 UTC m=+1291.129395260 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-scripts" not found Oct 04 07:31:12 crc kubenswrapper[4987]: E1004 07:31:12.074432 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 07:31:12 crc kubenswrapper[4987]: E1004 07:31:12.074637 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:14.074607101 +0000 UTC m=+1291.129505963 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-default-single-config-data" not found Oct 04 07:31:12 crc kubenswrapper[4987]: I1004 07:31:12.519167 4987 generic.go:334] "Generic (PLEG): container finished" podID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerID="b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f" exitCode=143 Oct 04 07:31:12 crc kubenswrapper[4987]: I1004 07:31:12.519266 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"462acbe8-3917-4ac3-afab-ca02a9ee0cb7","Type":"ContainerDied","Data":"b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f"} Oct 04 07:31:12 crc kubenswrapper[4987]: I1004 07:31:12.854863 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:12 crc kubenswrapper[4987]: I1004 07:31:12.990110 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bt68\" (UniqueName: \"kubernetes.io/projected/3bcaa891-328e-43c5-9ee7-748a5097cb0f-kube-api-access-4bt68\") pod \"3bcaa891-328e-43c5-9ee7-748a5097cb0f\" (UID: \"3bcaa891-328e-43c5-9ee7-748a5097cb0f\") " Oct 04 07:31:12 crc kubenswrapper[4987]: I1004 07:31:12.995277 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcaa891-328e-43c5-9ee7-748a5097cb0f-kube-api-access-4bt68" (OuterVolumeSpecName: "kube-api-access-4bt68") pod "3bcaa891-328e-43c5-9ee7-748a5097cb0f" (UID: "3bcaa891-328e-43c5-9ee7-748a5097cb0f"). InnerVolumeSpecName "kube-api-access-4bt68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:13 crc kubenswrapper[4987]: I1004 07:31:13.092404 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bt68\" (UniqueName: \"kubernetes.io/projected/3bcaa891-328e-43c5-9ee7-748a5097cb0f-kube-api-access-4bt68\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:13 crc kubenswrapper[4987]: I1004 07:31:13.527568 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" event={"ID":"3bcaa891-328e-43c5-9ee7-748a5097cb0f","Type":"ContainerDied","Data":"958a0ec940ec5f7bd8859c0a41ffbbeaf91b27b631ca8523188644f673372d5b"} Oct 04 07:31:13 crc kubenswrapper[4987]: I1004 07:31:13.527617 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="958a0ec940ec5f7bd8859c0a41ffbbeaf91b27b631ca8523188644f673372d5b" Oct 04 07:31:13 crc kubenswrapper[4987]: I1004 07:31:13.527741 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance9d75-account-delete-559xm" Oct 04 07:31:14 crc kubenswrapper[4987]: E1004 07:31:14.110787 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Oct 04 07:31:14 crc kubenswrapper[4987]: E1004 07:31:14.111256 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:18.111235566 +0000 UTC m=+1295.166134408 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-scripts" not found Oct 04 07:31:14 crc kubenswrapper[4987]: E1004 07:31:14.112576 4987 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Oct 04 07:31:14 crc kubenswrapper[4987]: E1004 07:31:14.112605 4987 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data podName:462acbe8-3917-4ac3-afab-ca02a9ee0cb7 nodeName:}" failed. No retries permitted until 2025-10-04 07:31:18.112595319 +0000 UTC m=+1295.167494161 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data") pod "glance-default-single-0" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7") : secret "glance-default-single-config-data" not found Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.091987 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.129853 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-logs\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.129938 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k69g\" (UniqueName: \"kubernetes.io/projected/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-kube-api-access-2k69g\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.129967 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-public-tls-certs\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.130028 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-httpd-run\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.130071 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-combined-ca-bundle\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.130129 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-internal-tls-certs\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.130238 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.130270 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.130312 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\" (UID: \"462acbe8-3917-4ac3-afab-ca02a9ee0cb7\") " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.131096 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.131495 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-logs" (OuterVolumeSpecName: "logs") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.140089 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.141475 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts" (OuterVolumeSpecName: "scripts") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.142872 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-kube-api-access-2k69g" (OuterVolumeSpecName: "kube-api-access-2k69g") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "kube-api-access-2k69g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.170003 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-kjbtd"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.181699 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-kjbtd"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.187799 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.192346 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.195061 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-9d75-account-create-fqxfz"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.203710 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data" (OuterVolumeSpecName: "config-data") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.205840 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "462acbe8-3917-4ac3-afab-ca02a9ee0cb7" (UID: "462acbe8-3917-4ac3-afab-ca02a9ee0cb7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.206976 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance9d75-account-delete-559xm"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.215009 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-9d75-account-create-fqxfz"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.223265 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance9d75-account-delete-559xm"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231768 4987 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231800 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231810 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231844 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231857 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231866 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k69g\" (UniqueName: \"kubernetes.io/projected/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-kube-api-access-2k69g\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231877 4987 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231887 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.231895 4987 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462acbe8-3917-4ac3-afab-ca02a9ee0cb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.245285 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.334031 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.553933 4987 generic.go:334] "Generic (PLEG): container finished" podID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerID="37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234" exitCode=0 Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.554007 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"462acbe8-3917-4ac3-afab-ca02a9ee0cb7","Type":"ContainerDied","Data":"37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234"} Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.554064 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"462acbe8-3917-4ac3-afab-ca02a9ee0cb7","Type":"ContainerDied","Data":"43a558703ac04027ede4d2c8e502ddc0a2617137bda077d16626a96401d40fd3"} Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.554091 4987 scope.go:117] "RemoveContainer" containerID="37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.554175 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.582744 4987 scope.go:117] "RemoveContainer" containerID="b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.604209 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.611254 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.611894 4987 scope.go:117] "RemoveContainer" containerID="37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234" Oct 04 07:31:15 crc kubenswrapper[4987]: E1004 07:31:15.612651 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234\": container with ID starting with 37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234 not found: ID does not exist" containerID="37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.612694 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234"} err="failed to get container status \"37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234\": rpc error: code = NotFound desc = could not find container \"37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234\": container with ID starting with 37bcd10936d4e879ced914ee59b5efd1706173887ba574984440cc92f1b47234 not found: ID does not exist" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.612723 4987 scope.go:117] "RemoveContainer" containerID="b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f" Oct 04 07:31:15 crc kubenswrapper[4987]: E1004 07:31:15.613221 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f\": container with ID starting with b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f not found: ID does not exist" containerID="b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f" Oct 04 07:31:15 crc kubenswrapper[4987]: I1004 07:31:15.613249 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f"} err="failed to get container status \"b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f\": rpc error: code = NotFound desc = could not find container \"b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f\": container with ID starting with b982d091fcb90efeb0d7551ccd2cb95e085ca040a758a30f64b450787abeb78f not found: ID does not exist" Oct 04 07:31:16 crc kubenswrapper[4987]: I1004 07:31:16.062534 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ed86d9-572f-4cc3-9a45-a299e9ce89d0" path="/var/lib/kubelet/pods/01ed86d9-572f-4cc3-9a45-a299e9ce89d0/volumes" Oct 04 07:31:16 crc kubenswrapper[4987]: I1004 07:31:16.063375 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bcaa891-328e-43c5-9ee7-748a5097cb0f" path="/var/lib/kubelet/pods/3bcaa891-328e-43c5-9ee7-748a5097cb0f/volumes" Oct 04 07:31:16 crc kubenswrapper[4987]: I1004 07:31:16.063886 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d181af4-dc7d-44e3-bb66-a06f0079ab7f" path="/var/lib/kubelet/pods/3d181af4-dc7d-44e3-bb66-a06f0079ab7f/volumes" Oct 04 07:31:16 crc kubenswrapper[4987]: I1004 07:31:16.064416 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" path="/var/lib/kubelet/pods/462acbe8-3917-4ac3-afab-ca02a9ee0cb7/volumes" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.075154 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-llv7g"] Oct 04 07:31:17 crc kubenswrapper[4987]: E1004 07:31:17.076113 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcaa891-328e-43c5-9ee7-748a5097cb0f" containerName="mariadb-account-delete" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.076137 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcaa891-328e-43c5-9ee7-748a5097cb0f" containerName="mariadb-account-delete" Oct 04 07:31:17 crc kubenswrapper[4987]: E1004 07:31:17.076158 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-log" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.076169 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-log" Oct 04 07:31:17 crc kubenswrapper[4987]: E1004 07:31:17.076185 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-httpd" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.076196 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-httpd" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.076405 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcaa891-328e-43c5-9ee7-748a5097cb0f" containerName="mariadb-account-delete" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.076432 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-httpd" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.076444 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="462acbe8-3917-4ac3-afab-ca02a9ee0cb7" containerName="glance-log" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.077164 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.090866 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-llv7g"] Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.165837 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z95s\" (UniqueName: \"kubernetes.io/projected/47de6379-1946-4012-afab-f15535d50880-kube-api-access-6z95s\") pod \"glance-db-create-llv7g\" (UID: \"47de6379-1946-4012-afab-f15535d50880\") " pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.267313 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z95s\" (UniqueName: \"kubernetes.io/projected/47de6379-1946-4012-afab-f15535d50880-kube-api-access-6z95s\") pod \"glance-db-create-llv7g\" (UID: \"47de6379-1946-4012-afab-f15535d50880\") " pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.288695 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z95s\" (UniqueName: \"kubernetes.io/projected/47de6379-1946-4012-afab-f15535d50880-kube-api-access-6z95s\") pod \"glance-db-create-llv7g\" (UID: \"47de6379-1946-4012-afab-f15535d50880\") " pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.400700 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:17 crc kubenswrapper[4987]: I1004 07:31:17.897891 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-llv7g"] Oct 04 07:31:18 crc kubenswrapper[4987]: I1004 07:31:18.594123 4987 generic.go:334] "Generic (PLEG): container finished" podID="47de6379-1946-4012-afab-f15535d50880" containerID="fc568d1505d3d44139b68fc46c9c52ef7bfbfb1f9cbe0fdd2bdc2915c1e8d852" exitCode=0 Oct 04 07:31:18 crc kubenswrapper[4987]: I1004 07:31:18.594223 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-llv7g" event={"ID":"47de6379-1946-4012-afab-f15535d50880","Type":"ContainerDied","Data":"fc568d1505d3d44139b68fc46c9c52ef7bfbfb1f9cbe0fdd2bdc2915c1e8d852"} Oct 04 07:31:18 crc kubenswrapper[4987]: I1004 07:31:18.594572 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-llv7g" event={"ID":"47de6379-1946-4012-afab-f15535d50880","Type":"ContainerStarted","Data":"f0960f1b4e794633b6305f6ef2900a73c19b735f89ff05cf6ba8a7ef3727e1f6"} Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.021071 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.114794 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z95s\" (UniqueName: \"kubernetes.io/projected/47de6379-1946-4012-afab-f15535d50880-kube-api-access-6z95s\") pod \"47de6379-1946-4012-afab-f15535d50880\" (UID: \"47de6379-1946-4012-afab-f15535d50880\") " Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.123451 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47de6379-1946-4012-afab-f15535d50880-kube-api-access-6z95s" (OuterVolumeSpecName: "kube-api-access-6z95s") pod "47de6379-1946-4012-afab-f15535d50880" (UID: "47de6379-1946-4012-afab-f15535d50880"). InnerVolumeSpecName "kube-api-access-6z95s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.217316 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z95s\" (UniqueName: \"kubernetes.io/projected/47de6379-1946-4012-afab-f15535d50880-kube-api-access-6z95s\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.612131 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-llv7g" event={"ID":"47de6379-1946-4012-afab-f15535d50880","Type":"ContainerDied","Data":"f0960f1b4e794633b6305f6ef2900a73c19b735f89ff05cf6ba8a7ef3727e1f6"} Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.612443 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-llv7g" Oct 04 07:31:20 crc kubenswrapper[4987]: I1004 07:31:20.612456 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0960f1b4e794633b6305f6ef2900a73c19b735f89ff05cf6ba8a7ef3727e1f6" Oct 04 07:31:21 crc kubenswrapper[4987]: E1004 07:31:21.002170 4987 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.122771 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-9e77-account-create-656md"] Oct 04 07:31:27 crc kubenswrapper[4987]: E1004 07:31:27.123886 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47de6379-1946-4012-afab-f15535d50880" containerName="mariadb-database-create" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.123902 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="47de6379-1946-4012-afab-f15535d50880" containerName="mariadb-database-create" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.124076 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="47de6379-1946-4012-afab-f15535d50880" containerName="mariadb-database-create" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.124646 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.132096 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.135002 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9e77-account-create-656md"] Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.243885 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fdp5\" (UniqueName: \"kubernetes.io/projected/cc2559bd-1713-4220-b9d2-41259f9b5d54-kube-api-access-8fdp5\") pod \"glance-9e77-account-create-656md\" (UID: \"cc2559bd-1713-4220-b9d2-41259f9b5d54\") " pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.345768 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fdp5\" (UniqueName: \"kubernetes.io/projected/cc2559bd-1713-4220-b9d2-41259f9b5d54-kube-api-access-8fdp5\") pod \"glance-9e77-account-create-656md\" (UID: \"cc2559bd-1713-4220-b9d2-41259f9b5d54\") " pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.368433 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fdp5\" (UniqueName: \"kubernetes.io/projected/cc2559bd-1713-4220-b9d2-41259f9b5d54-kube-api-access-8fdp5\") pod \"glance-9e77-account-create-656md\" (UID: \"cc2559bd-1713-4220-b9d2-41259f9b5d54\") " pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.459738 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:27 crc kubenswrapper[4987]: I1004 07:31:27.902258 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9e77-account-create-656md"] Oct 04 07:31:28 crc kubenswrapper[4987]: I1004 07:31:28.690851 4987 generic.go:334] "Generic (PLEG): container finished" podID="cc2559bd-1713-4220-b9d2-41259f9b5d54" containerID="24a15e9b45842d16127933029f731dda4bbf688478307e124e8df77cc136fdb3" exitCode=0 Oct 04 07:31:28 crc kubenswrapper[4987]: I1004 07:31:28.690946 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9e77-account-create-656md" event={"ID":"cc2559bd-1713-4220-b9d2-41259f9b5d54","Type":"ContainerDied","Data":"24a15e9b45842d16127933029f731dda4bbf688478307e124e8df77cc136fdb3"} Oct 04 07:31:28 crc kubenswrapper[4987]: I1004 07:31:28.691360 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9e77-account-create-656md" event={"ID":"cc2559bd-1713-4220-b9d2-41259f9b5d54","Type":"ContainerStarted","Data":"920c14e0e08ab3cd2773653fdd7bc81c1c8d946c213d15914d70dc39815dcecf"} Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.023176 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.094143 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fdp5\" (UniqueName: \"kubernetes.io/projected/cc2559bd-1713-4220-b9d2-41259f9b5d54-kube-api-access-8fdp5\") pod \"cc2559bd-1713-4220-b9d2-41259f9b5d54\" (UID: \"cc2559bd-1713-4220-b9d2-41259f9b5d54\") " Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.103835 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2559bd-1713-4220-b9d2-41259f9b5d54-kube-api-access-8fdp5" (OuterVolumeSpecName: "kube-api-access-8fdp5") pod "cc2559bd-1713-4220-b9d2-41259f9b5d54" (UID: "cc2559bd-1713-4220-b9d2-41259f9b5d54"). InnerVolumeSpecName "kube-api-access-8fdp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.195756 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fdp5\" (UniqueName: \"kubernetes.io/projected/cc2559bd-1713-4220-b9d2-41259f9b5d54-kube-api-access-8fdp5\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.711523 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9e77-account-create-656md" event={"ID":"cc2559bd-1713-4220-b9d2-41259f9b5d54","Type":"ContainerDied","Data":"920c14e0e08ab3cd2773653fdd7bc81c1c8d946c213d15914d70dc39815dcecf"} Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.711572 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e77-account-create-656md" Oct 04 07:31:30 crc kubenswrapper[4987]: I1004 07:31:30.711582 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="920c14e0e08ab3cd2773653fdd7bc81c1c8d946c213d15914d70dc39815dcecf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.195187 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-q4svf"] Oct 04 07:31:32 crc kubenswrapper[4987]: E1004 07:31:32.196966 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2559bd-1713-4220-b9d2-41259f9b5d54" containerName="mariadb-account-create" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.197094 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2559bd-1713-4220-b9d2-41259f9b5d54" containerName="mariadb-account-create" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.197376 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2559bd-1713-4220-b9d2-41259f9b5d54" containerName="mariadb-account-create" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.198289 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.201700 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.201920 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-p5cg5" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.219020 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-q4svf"] Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.327858 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fvvg\" (UniqueName: \"kubernetes.io/projected/d30bee49-7223-4652-9615-7155381ad95b-kube-api-access-4fvvg\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.327904 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-config-data\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.327959 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-db-sync-config-data\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.429342 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-db-sync-config-data\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.429467 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fvvg\" (UniqueName: \"kubernetes.io/projected/d30bee49-7223-4652-9615-7155381ad95b-kube-api-access-4fvvg\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.429503 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-config-data\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.449465 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-db-sync-config-data\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.452704 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-config-data\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.454680 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fvvg\" (UniqueName: \"kubernetes.io/projected/d30bee49-7223-4652-9615-7155381ad95b-kube-api-access-4fvvg\") pod \"glance-db-sync-q4svf\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:32 crc kubenswrapper[4987]: I1004 07:31:32.521435 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:33 crc kubenswrapper[4987]: I1004 07:31:33.005965 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-q4svf"] Oct 04 07:31:33 crc kubenswrapper[4987]: W1004 07:31:33.018491 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd30bee49_7223_4652_9615_7155381ad95b.slice/crio-6baf5894bf7f55196699d1501ff495e35bcb2f09ac0aee3422e20a2d6fa56195 WatchSource:0}: Error finding container 6baf5894bf7f55196699d1501ff495e35bcb2f09ac0aee3422e20a2d6fa56195: Status 404 returned error can't find the container with id 6baf5894bf7f55196699d1501ff495e35bcb2f09ac0aee3422e20a2d6fa56195 Oct 04 07:31:33 crc kubenswrapper[4987]: I1004 07:31:33.753144 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-q4svf" event={"ID":"d30bee49-7223-4652-9615-7155381ad95b","Type":"ContainerStarted","Data":"d9c1f7f2a5a8d9b56507bbba6bce078c17c6933e8777938d7deca9210e408f75"} Oct 04 07:31:33 crc kubenswrapper[4987]: I1004 07:31:33.753649 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-q4svf" event={"ID":"d30bee49-7223-4652-9615-7155381ad95b","Type":"ContainerStarted","Data":"6baf5894bf7f55196699d1501ff495e35bcb2f09ac0aee3422e20a2d6fa56195"} Oct 04 07:31:33 crc kubenswrapper[4987]: I1004 07:31:33.786922 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-q4svf" podStartSLOduration=1.786881003 podStartE2EDuration="1.786881003s" podCreationTimestamp="2025-10-04 07:31:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:31:33.784854769 +0000 UTC m=+1310.839753611" watchObservedRunningTime="2025-10-04 07:31:33.786881003 +0000 UTC m=+1310.841779885" Oct 04 07:31:36 crc kubenswrapper[4987]: I1004 07:31:36.778939 4987 generic.go:334] "Generic (PLEG): container finished" podID="d30bee49-7223-4652-9615-7155381ad95b" containerID="d9c1f7f2a5a8d9b56507bbba6bce078c17c6933e8777938d7deca9210e408f75" exitCode=0 Oct 04 07:31:36 crc kubenswrapper[4987]: I1004 07:31:36.779072 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-q4svf" event={"ID":"d30bee49-7223-4652-9615-7155381ad95b","Type":"ContainerDied","Data":"d9c1f7f2a5a8d9b56507bbba6bce078c17c6933e8777938d7deca9210e408f75"} Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.123403 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.229001 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-db-sync-config-data\") pod \"d30bee49-7223-4652-9615-7155381ad95b\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.230191 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-config-data\") pod \"d30bee49-7223-4652-9615-7155381ad95b\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.230308 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fvvg\" (UniqueName: \"kubernetes.io/projected/d30bee49-7223-4652-9615-7155381ad95b-kube-api-access-4fvvg\") pod \"d30bee49-7223-4652-9615-7155381ad95b\" (UID: \"d30bee49-7223-4652-9615-7155381ad95b\") " Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.238356 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d30bee49-7223-4652-9615-7155381ad95b-kube-api-access-4fvvg" (OuterVolumeSpecName: "kube-api-access-4fvvg") pod "d30bee49-7223-4652-9615-7155381ad95b" (UID: "d30bee49-7223-4652-9615-7155381ad95b"). InnerVolumeSpecName "kube-api-access-4fvvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.239172 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d30bee49-7223-4652-9615-7155381ad95b" (UID: "d30bee49-7223-4652-9615-7155381ad95b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.294171 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-config-data" (OuterVolumeSpecName: "config-data") pod "d30bee49-7223-4652-9615-7155381ad95b" (UID: "d30bee49-7223-4652-9615-7155381ad95b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.332824 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.332870 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fvvg\" (UniqueName: \"kubernetes.io/projected/d30bee49-7223-4652-9615-7155381ad95b-kube-api-access-4fvvg\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.332902 4987 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d30bee49-7223-4652-9615-7155381ad95b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.824465 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-q4svf" event={"ID":"d30bee49-7223-4652-9615-7155381ad95b","Type":"ContainerDied","Data":"6baf5894bf7f55196699d1501ff495e35bcb2f09ac0aee3422e20a2d6fa56195"} Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.824523 4987 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6baf5894bf7f55196699d1501ff495e35bcb2f09ac0aee3422e20a2d6fa56195" Oct 04 07:31:38 crc kubenswrapper[4987]: I1004 07:31:38.824799 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-q4svf" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.127422 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:40 crc kubenswrapper[4987]: E1004 07:31:40.128279 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30bee49-7223-4652-9615-7155381ad95b" containerName="glance-db-sync" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.128297 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30bee49-7223-4652-9615-7155381ad95b" containerName="glance-db-sync" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.128448 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30bee49-7223-4652-9615-7155381ad95b" containerName="glance-db-sync" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.129492 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.132453 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-p5cg5" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.133785 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.137753 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.140072 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:40 crc kubenswrapper[4987]: E1004 07:31:40.152179 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-wtljb lib-modules logs run scripts sys var-locks-brick], unattached volumes=[], failed to process volumes=[config-data dev etc-iscsi etc-nvme glance glance-cache httpd-run kube-api-access-wtljb lib-modules logs run scripts sys var-locks-brick]: context canceled" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="f141d961-1cab-4221-8ff8-16882b3ceae0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.168277 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.220784 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270377 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-sys\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270450 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-dev\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270498 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-config-data\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270567 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270586 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270606 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-logs\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270668 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270689 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-run\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270707 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-scripts\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270743 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270764 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270809 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtljb\" (UniqueName: \"kubernetes.io/projected/f141d961-1cab-4221-8ff8-16882b3ceae0-kube-api-access-wtljb\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270833 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.270878 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.271893 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.292913 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372402 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372463 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-run\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372484 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-scripts\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372512 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372563 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtljb\" (UniqueName: \"kubernetes.io/projected/f141d961-1cab-4221-8ff8-16882b3ceae0-kube-api-access-wtljb\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372598 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372643 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372662 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-sys\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372671 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372673 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-run\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372745 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-dev\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372690 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-dev\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372786 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372816 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-sys\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372882 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-config-data\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372952 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372952 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.372972 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.373010 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-logs\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.373276 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.373368 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.373497 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-logs\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.373538 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.378555 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-config-data\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.379509 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-scripts\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.391123 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtljb\" (UniqueName: \"kubernetes.io/projected/f141d961-1cab-4221-8ff8-16882b3ceae0-kube-api-access-wtljb\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.391938 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.535379 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.537249 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.540663 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.548379 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679438 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqrlg\" (UniqueName: \"kubernetes.io/projected/db4a3cef-191b-400e-9a54-e523c4d7e604-kube-api-access-rqrlg\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679510 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679551 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-dev\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679575 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679640 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-logs\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679671 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679704 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679725 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-scripts\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679756 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679790 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679827 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-sys\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679869 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-run\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679905 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-config-data\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.679938 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.781908 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqrlg\" (UniqueName: \"kubernetes.io/projected/db4a3cef-191b-400e-9a54-e523c4d7e604-kube-api-access-rqrlg\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.781979 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782017 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-dev\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782037 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782086 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-logs\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782111 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782134 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782155 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-scripts\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782175 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-dev\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782194 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782281 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782340 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-sys\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782376 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782445 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-run\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782386 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782553 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-sys\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782407 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-run\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782420 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782605 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782660 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-config-data\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782672 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782671 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782717 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.782838 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-logs\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.783200 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.788910 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-scripts\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.789764 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-config-data\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.809806 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqrlg\" (UniqueName: \"kubernetes.io/projected/db4a3cef-191b-400e-9a54-e523c4d7e604-kube-api-access-rqrlg\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.823471 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.826793 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.840612 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.853107 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:40 crc kubenswrapper[4987]: I1004 07:31:40.857945 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.016226 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-var-locks-brick\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.016893 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.016974 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-run" (OuterVolumeSpecName: "run") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.016949 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-run\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.017828 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-sys\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.017864 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtljb\" (UniqueName: \"kubernetes.io/projected/f141d961-1cab-4221-8ff8-16882b3ceae0-kube-api-access-wtljb\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.017896 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.017928 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-lib-modules\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.017959 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-dev\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018007 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-scripts\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018049 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-httpd-run\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018067 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-nvme\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018086 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-config-data\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018111 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-logs\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018143 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018163 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-iscsi\") pod \"f141d961-1cab-4221-8ff8-16882b3ceae0\" (UID: \"f141d961-1cab-4221-8ff8-16882b3ceae0\") " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018546 4987 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018561 4987 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018597 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.018968 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-sys" (OuterVolumeSpecName: "sys") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.019058 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-dev" (OuterVolumeSpecName: "dev") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.019082 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.019305 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.019894 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-logs" (OuterVolumeSpecName: "logs") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.019931 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.024055 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance-cache") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.024068 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-scripts" (OuterVolumeSpecName: "scripts") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.024758 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-config-data" (OuterVolumeSpecName: "config-data") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.024869 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f141d961-1cab-4221-8ff8-16882b3ceae0-kube-api-access-wtljb" (OuterVolumeSpecName: "kube-api-access-wtljb") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "kube-api-access-wtljb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.035113 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f141d961-1cab-4221-8ff8-16882b3ceae0" (UID: "f141d961-1cab-4221-8ff8-16882b3ceae0"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120053 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120113 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120126 4987 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120137 4987 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-sys\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120153 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtljb\" (UniqueName: \"kubernetes.io/projected/f141d961-1cab-4221-8ff8-16882b3ceae0-kube-api-access-wtljb\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120168 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120177 4987 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120185 4987 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-dev\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120193 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120200 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f141d961-1cab-4221-8ff8-16882b3ceae0-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120208 4987 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f141d961-1cab-4221-8ff8-16882b3ceae0-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.120216 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f141d961-1cab-4221-8ff8-16882b3ceae0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.133506 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.134315 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.221433 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.221473 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.355005 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.571013 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.852580 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerStarted","Data":"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028"} Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.853536 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerStarted","Data":"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1"} Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.853575 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerStarted","Data":"7337405f8054b15ccefdaf10f9d8e7b2ec81b193bfb5111165e222a4084e64b4"} Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.852616 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.911313 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.922489 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.953015 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.955144 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.959067 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Oct 04 07:31:41 crc kubenswrapper[4987]: I1004 07:31:41.983481 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.064069 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f141d961-1cab-4221-8ff8-16882b3ceae0" path="/var/lib/kubelet/pods/f141d961-1cab-4221-8ff8-16882b3ceae0/volumes" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.135691 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-run\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.135759 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.135792 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-scripts\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136056 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-dev\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136169 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136211 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-sys\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136307 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136337 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136374 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-logs\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136412 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136457 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136546 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136747 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnvwc\" (UniqueName: \"kubernetes.io/projected/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-kube-api-access-rnvwc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.136816 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-config-data\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.238753 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnvwc\" (UniqueName: \"kubernetes.io/projected/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-kube-api-access-rnvwc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.238830 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-config-data\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.238862 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-run\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.238928 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.238955 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-scripts\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239026 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-dev\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239059 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239051 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-run\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239086 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-sys\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239140 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-sys\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239270 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239311 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239373 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-logs\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239431 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239460 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.239561 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.240382 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.240764 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.240827 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-dev\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.240865 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.241216 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.241308 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.240869 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.241680 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-logs\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.241779 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.249228 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-scripts\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.249394 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-config-data\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.261109 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnvwc\" (UniqueName: \"kubernetes.io/projected/6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26-kube-api-access-rnvwc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.261442 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.261756 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26\") " pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.274347 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.703763 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.704334 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.757846 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Oct 04 07:31:42 crc kubenswrapper[4987]: W1004 07:31:42.767757 4987 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d400a6b_b3eb_4cfb_bb07_5ad4e6522a26.slice/crio-70c51035dfc7db38996a83a4e5a5d29b8baa15b7680fbae3b4c1c01514ef3682 WatchSource:0}: Error finding container 70c51035dfc7db38996a83a4e5a5d29b8baa15b7680fbae3b4c1c01514ef3682: Status 404 returned error can't find the container with id 70c51035dfc7db38996a83a4e5a5d29b8baa15b7680fbae3b4c1c01514ef3682 Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.863433 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerStarted","Data":"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a"} Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.863760 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-log" containerID="cri-o://7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" gracePeriod=30 Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.864036 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-api" containerID="cri-o://bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" gracePeriod=30 Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.864094 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-httpd" containerID="cri-o://66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" gracePeriod=30 Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.866682 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26","Type":"ContainerStarted","Data":"70c51035dfc7db38996a83a4e5a5d29b8baa15b7680fbae3b4c1c01514ef3682"} Oct 04 07:31:42 crc kubenswrapper[4987]: I1004 07:31:42.901453 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.901425986 podStartE2EDuration="3.901425986s" podCreationTimestamp="2025-10-04 07:31:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:31:42.896257742 +0000 UTC m=+1319.951156584" watchObservedRunningTime="2025-10-04 07:31:42.901425986 +0000 UTC m=+1319.956324828" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.359595 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.461499 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-logs\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.461952 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-sys\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.461999 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-iscsi\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462023 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-dev\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462078 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-config-data\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462099 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-nvme\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462163 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462215 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-httpd-run\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462233 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462261 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-run\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462291 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-var-locks-brick\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462314 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-scripts\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462348 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqrlg\" (UniqueName: \"kubernetes.io/projected/db4a3cef-191b-400e-9a54-e523c4d7e604-kube-api-access-rqrlg\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462399 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-lib-modules\") pod \"db4a3cef-191b-400e-9a54-e523c4d7e604\" (UID: \"db4a3cef-191b-400e-9a54-e523c4d7e604\") " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462201 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-logs" (OuterVolumeSpecName: "logs") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462768 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462787 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.462946 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.463000 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-sys" (OuterVolumeSpecName: "sys") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.463049 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-run" (OuterVolumeSpecName: "run") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.463452 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.463611 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-dev" (OuterVolumeSpecName: "dev") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.464191 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.468399 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.468376 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-scripts" (OuterVolumeSpecName: "scripts") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.469540 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance-cache") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.471107 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4a3cef-191b-400e-9a54-e523c4d7e604-kube-api-access-rqrlg" (OuterVolumeSpecName: "kube-api-access-rqrlg") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "kube-api-access-rqrlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.537374 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-config-data" (OuterVolumeSpecName: "config-data") pod "db4a3cef-191b-400e-9a54-e523c4d7e604" (UID: "db4a3cef-191b-400e-9a54-e523c4d7e604"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.564568 4987 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-logs\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.564601 4987 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-sys\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.564613 4987 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-iscsi\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565149 4987 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-dev\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565175 4987 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565187 4987 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-etc-nvme\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565224 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565236 4987 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/db4a3cef-191b-400e-9a54-e523c4d7e604-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565250 4987 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565260 4987 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-run\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565270 4987 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-var-locks-brick\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565282 4987 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/db4a3cef-191b-400e-9a54-e523c4d7e604-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565292 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqrlg\" (UniqueName: \"kubernetes.io/projected/db4a3cef-191b-400e-9a54-e523c4d7e604-kube-api-access-rqrlg\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.565301 4987 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/db4a3cef-191b-400e-9a54-e523c4d7e604-lib-modules\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.581135 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.584176 4987 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.666657 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.666687 4987 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.876976 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26","Type":"ContainerStarted","Data":"70253647a3a0deb61802348b7e8afe2d03ea3d179e023ed9376c980ea72c50e4"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.877040 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26","Type":"ContainerStarted","Data":"f2b7a0c45a78cd3a0e1ec5e84c379a2cfdfb0e63b9bd3a301add315fc37b1e31"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.877050 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26","Type":"ContainerStarted","Data":"5ff71caaaf158a2c88fd18b55428a1b171c27599053e679812f258a8b55a15e5"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880575 4987 generic.go:334] "Generic (PLEG): container finished" podID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerID="bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" exitCode=143 Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880641 4987 generic.go:334] "Generic (PLEG): container finished" podID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerID="66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" exitCode=143 Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880656 4987 generic.go:334] "Generic (PLEG): container finished" podID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerID="7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" exitCode=143 Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880682 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerDied","Data":"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880685 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880708 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerDied","Data":"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880722 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerDied","Data":"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880737 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"db4a3cef-191b-400e-9a54-e523c4d7e604","Type":"ContainerDied","Data":"7337405f8054b15ccefdaf10f9d8e7b2ec81b193bfb5111165e222a4084e64b4"} Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.880759 4987 scope.go:117] "RemoveContainer" containerID="bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.927823 4987 scope.go:117] "RemoveContainer" containerID="66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.934004 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.933976319 podStartE2EDuration="2.933976319s" podCreationTimestamp="2025-10-04 07:31:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:31:43.906869329 +0000 UTC m=+1320.961768171" watchObservedRunningTime="2025-10-04 07:31:43.933976319 +0000 UTC m=+1320.988875161" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.956012 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.960610 4987 scope.go:117] "RemoveContainer" containerID="7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.975436 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.988797 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:43 crc kubenswrapper[4987]: E1004 07:31:43.991016 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-httpd" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.991103 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-httpd" Oct 04 07:31:43 crc kubenswrapper[4987]: E1004 07:31:43.991119 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-log" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.991128 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-log" Oct 04 07:31:43 crc kubenswrapper[4987]: E1004 07:31:43.991165 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-api" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.991172 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-api" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.991331 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-httpd" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.991347 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-api" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.991355 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" containerName="glance-log" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.992441 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.993845 4987 scope.go:117] "RemoveContainer" containerID="bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" Oct 04 07:31:43 crc kubenswrapper[4987]: E1004 07:31:43.997737 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": container with ID starting with bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a not found: ID does not exist" containerID="bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.997774 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a"} err="failed to get container status \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": rpc error: code = NotFound desc = could not find container \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": container with ID starting with bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a not found: ID does not exist" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.997807 4987 scope.go:117] "RemoveContainer" containerID="66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" Oct 04 07:31:43 crc kubenswrapper[4987]: E1004 07:31:43.998303 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": container with ID starting with 66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028 not found: ID does not exist" containerID="66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.998346 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028"} err="failed to get container status \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": rpc error: code = NotFound desc = could not find container \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": container with ID starting with 66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028 not found: ID does not exist" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.998382 4987 scope.go:117] "RemoveContainer" containerID="7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.998566 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Oct 04 07:31:43 crc kubenswrapper[4987]: E1004 07:31:43.998848 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": container with ID starting with 7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1 not found: ID does not exist" containerID="7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.998870 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1"} err="failed to get container status \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": rpc error: code = NotFound desc = could not find container \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": container with ID starting with 7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1 not found: ID does not exist" Oct 04 07:31:43 crc kubenswrapper[4987]: I1004 07:31:43.998886 4987 scope.go:117] "RemoveContainer" containerID="bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.003733 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a"} err="failed to get container status \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": rpc error: code = NotFound desc = could not find container \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": container with ID starting with bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a not found: ID does not exist" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.003781 4987 scope.go:117] "RemoveContainer" containerID="66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.010113 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028"} err="failed to get container status \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": rpc error: code = NotFound desc = could not find container \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": container with ID starting with 66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028 not found: ID does not exist" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.010170 4987 scope.go:117] "RemoveContainer" containerID="7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.015148 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1"} err="failed to get container status \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": rpc error: code = NotFound desc = could not find container \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": container with ID starting with 7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1 not found: ID does not exist" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.015209 4987 scope.go:117] "RemoveContainer" containerID="bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.016036 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.018507 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a"} err="failed to get container status \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": rpc error: code = NotFound desc = could not find container \"bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a\": container with ID starting with bd949941a2f35180693e33345be637919cf67b6b7ac896d3d2c97f924eab496a not found: ID does not exist" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.018592 4987 scope.go:117] "RemoveContainer" containerID="66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.026908 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028"} err="failed to get container status \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": rpc error: code = NotFound desc = could not find container \"66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028\": container with ID starting with 66f2497bc06526bf8aef89222142c0272f7d1ad034a2bea30e6eaa974154a028 not found: ID does not exist" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.026965 4987 scope.go:117] "RemoveContainer" containerID="7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.038864 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1"} err="failed to get container status \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": rpc error: code = NotFound desc = could not find container \"7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1\": container with ID starting with 7592231b24a89985621b6d2ac93395583be476052c499d2715a88a1cea7a2bb1 not found: ID does not exist" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.076150 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.077716 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4a3cef-191b-400e-9a54-e523c4d7e604" path="/var/lib/kubelet/pods/db4a3cef-191b-400e-9a54-e523c4d7e604/volumes" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177495 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177542 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b0984a-9602-4187-b515-f862cf9f2b4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177575 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177598 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b0984a-9602-4187-b515-f862cf9f2b4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177642 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177659 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b0984a-9602-4187-b515-f862cf9f2b4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177691 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-dev\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177709 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-sys\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177733 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87r87\" (UniqueName: \"kubernetes.io/projected/d6b0984a-9602-4187-b515-f862cf9f2b4b-kube-api-access-87r87\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177763 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177787 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177808 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177831 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.177860 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b0984a-9602-4187-b515-f862cf9f2b4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.178220 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.201582 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279812 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-sys\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279890 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87r87\" (UniqueName: \"kubernetes.io/projected/d6b0984a-9602-4187-b515-f862cf9f2b4b-kube-api-access-87r87\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279925 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279949 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279972 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279985 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-sys\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.279998 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b0984a-9602-4187-b515-f862cf9f2b4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280167 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280186 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b0984a-9602-4187-b515-f862cf9f2b4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280199 4987 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280605 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280815 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280796 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280725 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280979 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b0984a-9602-4187-b515-f862cf9f2b4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.280859 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281198 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b0984a-9602-4187-b515-f862cf9f2b4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281243 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281329 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-dev\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281451 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281460 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6b0984a-9602-4187-b515-f862cf9f2b4b-dev\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281648 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b0984a-9602-4187-b515-f862cf9f2b4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.281737 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b0984a-9602-4187-b515-f862cf9f2b4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.286355 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b0984a-9602-4187-b515-f862cf9f2b4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.293831 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b0984a-9602-4187-b515-f862cf9f2b4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.297495 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87r87\" (UniqueName: \"kubernetes.io/projected/d6b0984a-9602-4187-b515-f862cf9f2b4b-kube-api-access-87r87\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.301610 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b0984a-9602-4187-b515-f862cf9f2b4b\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.320556 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.737564 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Oct 04 07:31:44 crc kubenswrapper[4987]: I1004 07:31:44.890084 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d6b0984a-9602-4187-b515-f862cf9f2b4b","Type":"ContainerStarted","Data":"2c3abdf084650fd4ccf025a7d32347b553bfa5e3c2495f7c60264f5ce358401a"} Oct 04 07:31:45 crc kubenswrapper[4987]: I1004 07:31:45.903560 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d6b0984a-9602-4187-b515-f862cf9f2b4b","Type":"ContainerStarted","Data":"d3b020e41233b77d3d8b4fcb1aac008dee8891e904cd4a880cb03bd4b9864f76"} Oct 04 07:31:45 crc kubenswrapper[4987]: I1004 07:31:45.904388 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d6b0984a-9602-4187-b515-f862cf9f2b4b","Type":"ContainerStarted","Data":"79833deea79cd98b0f0659630c3de24a32d5b20dac7d4fc12e6774afdd141a86"} Oct 04 07:31:45 crc kubenswrapper[4987]: I1004 07:31:45.904404 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d6b0984a-9602-4187-b515-f862cf9f2b4b","Type":"ContainerStarted","Data":"e52d709df388463901a58a072b841c5f6508abae6f026e93800f35abb5d048a8"} Oct 04 07:31:45 crc kubenswrapper[4987]: I1004 07:31:45.929176 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.929149799 podStartE2EDuration="2.929149799s" podCreationTimestamp="2025-10-04 07:31:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:31:45.926530305 +0000 UTC m=+1322.981429157" watchObservedRunningTime="2025-10-04 07:31:45.929149799 +0000 UTC m=+1322.984048641" Oct 04 07:31:52 crc kubenswrapper[4987]: I1004 07:31:52.274948 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:52 crc kubenswrapper[4987]: I1004 07:31:52.275902 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:52 crc kubenswrapper[4987]: I1004 07:31:52.275928 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:52 crc kubenswrapper[4987]: I1004 07:31:52.302962 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:52 crc kubenswrapper[4987]: I1004 07:31:52.303055 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:52 crc kubenswrapper[4987]: I1004 07:31:52.315299 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:53 crc kubenswrapper[4987]: I1004 07:31:53.020092 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:53 crc kubenswrapper[4987]: I1004 07:31:53.021574 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:53 crc kubenswrapper[4987]: I1004 07:31:53.021780 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:53 crc kubenswrapper[4987]: I1004 07:31:53.033184 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:53 crc kubenswrapper[4987]: I1004 07:31:53.033762 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:53 crc kubenswrapper[4987]: I1004 07:31:53.036170 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Oct 04 07:31:54 crc kubenswrapper[4987]: I1004 07:31:54.321228 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:54 crc kubenswrapper[4987]: I1004 07:31:54.321299 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:54 crc kubenswrapper[4987]: I1004 07:31:54.321309 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:54 crc kubenswrapper[4987]: I1004 07:31:54.354473 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:54 crc kubenswrapper[4987]: I1004 07:31:54.358406 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:54 crc kubenswrapper[4987]: I1004 07:31:54.370281 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:55 crc kubenswrapper[4987]: I1004 07:31:55.043147 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:55 crc kubenswrapper[4987]: I1004 07:31:55.043191 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:55 crc kubenswrapper[4987]: I1004 07:31:55.043204 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:55 crc kubenswrapper[4987]: I1004 07:31:55.061089 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:55 crc kubenswrapper[4987]: I1004 07:31:55.062306 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:31:55 crc kubenswrapper[4987]: I1004 07:31:55.065913 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Oct 04 07:32:12 crc kubenswrapper[4987]: I1004 07:32:12.703103 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:32:12 crc kubenswrapper[4987]: I1004 07:32:12.703818 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:32:42 crc kubenswrapper[4987]: I1004 07:32:42.703357 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:32:42 crc kubenswrapper[4987]: I1004 07:32:42.704143 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:32:42 crc kubenswrapper[4987]: I1004 07:32:42.704211 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:32:42 crc kubenswrapper[4987]: I1004 07:32:42.705194 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"901e84e5b40d7f1df5e525731897cdcc9e887088810c1590e56d6e773c746b71"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:32:42 crc kubenswrapper[4987]: I1004 07:32:42.705268 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://901e84e5b40d7f1df5e525731897cdcc9e887088810c1590e56d6e773c746b71" gracePeriod=600 Oct 04 07:32:43 crc kubenswrapper[4987]: I1004 07:32:43.496741 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"901e84e5b40d7f1df5e525731897cdcc9e887088810c1590e56d6e773c746b71"} Oct 04 07:32:43 crc kubenswrapper[4987]: I1004 07:32:43.496772 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="901e84e5b40d7f1df5e525731897cdcc9e887088810c1590e56d6e773c746b71" exitCode=0 Oct 04 07:32:43 crc kubenswrapper[4987]: I1004 07:32:43.497580 4987 scope.go:117] "RemoveContainer" containerID="9851ab3918efc181099219aac8c4cb4de265cdf08bb6a6a2096a913ebd0e22d6" Oct 04 07:32:43 crc kubenswrapper[4987]: I1004 07:32:43.497689 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d"} Oct 04 07:33:24 crc kubenswrapper[4987]: I1004 07:33:24.866781 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q472h"] Oct 04 07:33:24 crc kubenswrapper[4987]: I1004 07:33:24.869731 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:24 crc kubenswrapper[4987]: I1004 07:33:24.877203 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q472h"] Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.003448 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-utilities\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.003512 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtdsb\" (UniqueName: \"kubernetes.io/projected/e0887015-d6fc-4def-9704-4d35ae41ec65-kube-api-access-qtdsb\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.003712 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-catalog-content\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.105983 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-utilities\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.106045 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtdsb\" (UniqueName: \"kubernetes.io/projected/e0887015-d6fc-4def-9704-4d35ae41ec65-kube-api-access-qtdsb\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.106076 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-catalog-content\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.106589 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-utilities\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.106724 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-catalog-content\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.143020 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtdsb\" (UniqueName: \"kubernetes.io/projected/e0887015-d6fc-4def-9704-4d35ae41ec65-kube-api-access-qtdsb\") pod \"community-operators-q472h\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.230020 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.759891 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q472h"] Oct 04 07:33:25 crc kubenswrapper[4987]: I1004 07:33:25.944555 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q472h" event={"ID":"e0887015-d6fc-4def-9704-4d35ae41ec65","Type":"ContainerStarted","Data":"21fa3687ecb615a1775337c562785582827c2e58526978041ef619e3ccedd6a3"} Oct 04 07:33:26 crc kubenswrapper[4987]: I1004 07:33:26.956849 4987 generic.go:334] "Generic (PLEG): container finished" podID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerID="f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b" exitCode=0 Oct 04 07:33:26 crc kubenswrapper[4987]: I1004 07:33:26.956965 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q472h" event={"ID":"e0887015-d6fc-4def-9704-4d35ae41ec65","Type":"ContainerDied","Data":"f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b"} Oct 04 07:33:26 crc kubenswrapper[4987]: I1004 07:33:26.959485 4987 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 07:33:28 crc kubenswrapper[4987]: I1004 07:33:28.977863 4987 generic.go:334] "Generic (PLEG): container finished" podID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerID="47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c" exitCode=0 Oct 04 07:33:28 crc kubenswrapper[4987]: I1004 07:33:28.977975 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q472h" event={"ID":"e0887015-d6fc-4def-9704-4d35ae41ec65","Type":"ContainerDied","Data":"47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c"} Oct 04 07:33:31 crc kubenswrapper[4987]: I1004 07:33:31.000479 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q472h" event={"ID":"e0887015-d6fc-4def-9704-4d35ae41ec65","Type":"ContainerStarted","Data":"ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de"} Oct 04 07:33:31 crc kubenswrapper[4987]: I1004 07:33:31.020872 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q472h" podStartSLOduration=4.049181866 podStartE2EDuration="7.020850666s" podCreationTimestamp="2025-10-04 07:33:24 +0000 UTC" firstStartedPulling="2025-10-04 07:33:26.959173253 +0000 UTC m=+1424.014072105" lastFinishedPulling="2025-10-04 07:33:29.930842073 +0000 UTC m=+1426.985740905" observedRunningTime="2025-10-04 07:33:31.019527355 +0000 UTC m=+1428.074426217" watchObservedRunningTime="2025-10-04 07:33:31.020850666 +0000 UTC m=+1428.075749508" Oct 04 07:33:35 crc kubenswrapper[4987]: I1004 07:33:35.230249 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:35 crc kubenswrapper[4987]: I1004 07:33:35.231383 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:35 crc kubenswrapper[4987]: I1004 07:33:35.273800 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:36 crc kubenswrapper[4987]: I1004 07:33:36.098138 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:36 crc kubenswrapper[4987]: I1004 07:33:36.148959 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q472h"] Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.064838 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q472h" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="registry-server" containerID="cri-o://ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de" gracePeriod=2 Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.500319 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.647984 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtdsb\" (UniqueName: \"kubernetes.io/projected/e0887015-d6fc-4def-9704-4d35ae41ec65-kube-api-access-qtdsb\") pod \"e0887015-d6fc-4def-9704-4d35ae41ec65\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.648654 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-utilities\") pod \"e0887015-d6fc-4def-9704-4d35ae41ec65\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.648961 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-catalog-content\") pod \"e0887015-d6fc-4def-9704-4d35ae41ec65\" (UID: \"e0887015-d6fc-4def-9704-4d35ae41ec65\") " Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.650160 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-utilities" (OuterVolumeSpecName: "utilities") pod "e0887015-d6fc-4def-9704-4d35ae41ec65" (UID: "e0887015-d6fc-4def-9704-4d35ae41ec65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.657368 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0887015-d6fc-4def-9704-4d35ae41ec65-kube-api-access-qtdsb" (OuterVolumeSpecName: "kube-api-access-qtdsb") pod "e0887015-d6fc-4def-9704-4d35ae41ec65" (UID: "e0887015-d6fc-4def-9704-4d35ae41ec65"). InnerVolumeSpecName "kube-api-access-qtdsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.701128 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0887015-d6fc-4def-9704-4d35ae41ec65" (UID: "e0887015-d6fc-4def-9704-4d35ae41ec65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.751353 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.751400 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtdsb\" (UniqueName: \"kubernetes.io/projected/e0887015-d6fc-4def-9704-4d35ae41ec65-kube-api-access-qtdsb\") on node \"crc\" DevicePath \"\"" Oct 04 07:33:38 crc kubenswrapper[4987]: I1004 07:33:38.751412 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0887015-d6fc-4def-9704-4d35ae41ec65-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.074500 4987 generic.go:334] "Generic (PLEG): container finished" podID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerID="ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de" exitCode=0 Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.074548 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q472h" event={"ID":"e0887015-d6fc-4def-9704-4d35ae41ec65","Type":"ContainerDied","Data":"ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de"} Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.074664 4987 scope.go:117] "RemoveContainer" containerID="ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.074819 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q472h" event={"ID":"e0887015-d6fc-4def-9704-4d35ae41ec65","Type":"ContainerDied","Data":"21fa3687ecb615a1775337c562785582827c2e58526978041ef619e3ccedd6a3"} Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.075980 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q472h" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.105307 4987 scope.go:117] "RemoveContainer" containerID="47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.116830 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q472h"] Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.123873 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q472h"] Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.143043 4987 scope.go:117] "RemoveContainer" containerID="f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.171817 4987 scope.go:117] "RemoveContainer" containerID="ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de" Oct 04 07:33:39 crc kubenswrapper[4987]: E1004 07:33:39.172540 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de\": container with ID starting with ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de not found: ID does not exist" containerID="ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.172600 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de"} err="failed to get container status \"ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de\": rpc error: code = NotFound desc = could not find container \"ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de\": container with ID starting with ed7089ee604f93503c51d6780bd76b4cd4f90f2aeca02bd6388b51e977d668de not found: ID does not exist" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.172679 4987 scope.go:117] "RemoveContainer" containerID="47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c" Oct 04 07:33:39 crc kubenswrapper[4987]: E1004 07:33:39.173039 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c\": container with ID starting with 47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c not found: ID does not exist" containerID="47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.173058 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c"} err="failed to get container status \"47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c\": rpc error: code = NotFound desc = could not find container \"47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c\": container with ID starting with 47577bff7f740e9d0f66b604b89699d15769fdbd02f80957f9ee7729af0a443c not found: ID does not exist" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.173073 4987 scope.go:117] "RemoveContainer" containerID="f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b" Oct 04 07:33:39 crc kubenswrapper[4987]: E1004 07:33:39.173351 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b\": container with ID starting with f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b not found: ID does not exist" containerID="f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b" Oct 04 07:33:39 crc kubenswrapper[4987]: I1004 07:33:39.173367 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b"} err="failed to get container status \"f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b\": rpc error: code = NotFound desc = could not find container \"f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b\": container with ID starting with f31b7191a58bac4970aaec9c40d2a34ebbe07965cefbd48010105e108e51cd8b not found: ID does not exist" Oct 04 07:33:40 crc kubenswrapper[4987]: I1004 07:33:40.064391 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" path="/var/lib/kubelet/pods/e0887015-d6fc-4def-9704-4d35ae41ec65/volumes" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.483486 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sxb2r"] Oct 04 07:34:03 crc kubenswrapper[4987]: E1004 07:34:03.484793 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="extract-utilities" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.484812 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="extract-utilities" Oct 04 07:34:03 crc kubenswrapper[4987]: E1004 07:34:03.484839 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="extract-content" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.484846 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="extract-content" Oct 04 07:34:03 crc kubenswrapper[4987]: E1004 07:34:03.484870 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="registry-server" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.484877 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="registry-server" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.485043 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0887015-d6fc-4def-9704-4d35ae41ec65" containerName="registry-server" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.486248 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.507306 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sxb2r"] Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.609778 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-utilities\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.609854 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-catalog-content\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.609901 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcrhw\" (UniqueName: \"kubernetes.io/projected/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-kube-api-access-wcrhw\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.712406 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-catalog-content\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.712547 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcrhw\" (UniqueName: \"kubernetes.io/projected/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-kube-api-access-wcrhw\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.712709 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-utilities\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.713201 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-catalog-content\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.713253 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-utilities\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.747109 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcrhw\" (UniqueName: \"kubernetes.io/projected/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-kube-api-access-wcrhw\") pod \"redhat-operators-sxb2r\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:03 crc kubenswrapper[4987]: I1004 07:34:03.813816 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:04 crc kubenswrapper[4987]: I1004 07:34:04.137715 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sxb2r"] Oct 04 07:34:04 crc kubenswrapper[4987]: I1004 07:34:04.286966 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerStarted","Data":"4bf5d4fcda3e568fef9060d2cf4b01c85c30d5d23230900329353f0ab122bc3c"} Oct 04 07:34:05 crc kubenswrapper[4987]: I1004 07:34:05.295859 4987 generic.go:334] "Generic (PLEG): container finished" podID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerID="a25c25285e05971d23987ac5467f675bf4a5ac5b9f2f873407ed8885edfc6f33" exitCode=0 Oct 04 07:34:05 crc kubenswrapper[4987]: I1004 07:34:05.295972 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerDied","Data":"a25c25285e05971d23987ac5467f675bf4a5ac5b9f2f873407ed8885edfc6f33"} Oct 04 07:34:06 crc kubenswrapper[4987]: I1004 07:34:06.306151 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerStarted","Data":"b021bd8051fc2a1ae99fcb55af1e750d94af08fd500642c244e23331d7abbd6c"} Oct 04 07:34:07 crc kubenswrapper[4987]: I1004 07:34:07.317895 4987 generic.go:334] "Generic (PLEG): container finished" podID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerID="b021bd8051fc2a1ae99fcb55af1e750d94af08fd500642c244e23331d7abbd6c" exitCode=0 Oct 04 07:34:07 crc kubenswrapper[4987]: I1004 07:34:07.317974 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerDied","Data":"b021bd8051fc2a1ae99fcb55af1e750d94af08fd500642c244e23331d7abbd6c"} Oct 04 07:34:08 crc kubenswrapper[4987]: I1004 07:34:08.329113 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerStarted","Data":"9667d84351c5602e3b92bbdd86f1472fa73fcefd4992586a914ed459dcf970b0"} Oct 04 07:34:13 crc kubenswrapper[4987]: I1004 07:34:13.814095 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:13 crc kubenswrapper[4987]: I1004 07:34:13.815046 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:13 crc kubenswrapper[4987]: I1004 07:34:13.870973 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:13 crc kubenswrapper[4987]: I1004 07:34:13.900932 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sxb2r" podStartSLOduration=8.488655773 podStartE2EDuration="10.900908157s" podCreationTimestamp="2025-10-04 07:34:03 +0000 UTC" firstStartedPulling="2025-10-04 07:34:05.297487844 +0000 UTC m=+1462.352386686" lastFinishedPulling="2025-10-04 07:34:07.709740228 +0000 UTC m=+1464.764639070" observedRunningTime="2025-10-04 07:34:08.352222482 +0000 UTC m=+1465.407121334" watchObservedRunningTime="2025-10-04 07:34:13.900908157 +0000 UTC m=+1470.955806999" Oct 04 07:34:14 crc kubenswrapper[4987]: I1004 07:34:14.429470 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:14 crc kubenswrapper[4987]: I1004 07:34:14.484564 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sxb2r"] Oct 04 07:34:16 crc kubenswrapper[4987]: I1004 07:34:16.393961 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sxb2r" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="registry-server" containerID="cri-o://9667d84351c5602e3b92bbdd86f1472fa73fcefd4992586a914ed459dcf970b0" gracePeriod=2 Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.404931 4987 generic.go:334] "Generic (PLEG): container finished" podID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerID="9667d84351c5602e3b92bbdd86f1472fa73fcefd4992586a914ed459dcf970b0" exitCode=0 Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.405036 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerDied","Data":"9667d84351c5602e3b92bbdd86f1472fa73fcefd4992586a914ed459dcf970b0"} Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.902358 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.966363 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcrhw\" (UniqueName: \"kubernetes.io/projected/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-kube-api-access-wcrhw\") pod \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.966433 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-catalog-content\") pod \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.966469 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-utilities\") pod \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\" (UID: \"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a\") " Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.967548 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-utilities" (OuterVolumeSpecName: "utilities") pod "5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" (UID: "5f539ae6-9dc4-4b26-b7bc-d4052346ec0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:34:17 crc kubenswrapper[4987]: I1004 07:34:17.972943 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-kube-api-access-wcrhw" (OuterVolumeSpecName: "kube-api-access-wcrhw") pod "5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" (UID: "5f539ae6-9dc4-4b26-b7bc-d4052346ec0a"). InnerVolumeSpecName "kube-api-access-wcrhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.048773 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" (UID: "5f539ae6-9dc4-4b26-b7bc-d4052346ec0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.069002 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcrhw\" (UniqueName: \"kubernetes.io/projected/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-kube-api-access-wcrhw\") on node \"crc\" DevicePath \"\"" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.069046 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.069057 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.420057 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sxb2r" event={"ID":"5f539ae6-9dc4-4b26-b7bc-d4052346ec0a","Type":"ContainerDied","Data":"4bf5d4fcda3e568fef9060d2cf4b01c85c30d5d23230900329353f0ab122bc3c"} Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.420142 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sxb2r" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.420152 4987 scope.go:117] "RemoveContainer" containerID="9667d84351c5602e3b92bbdd86f1472fa73fcefd4992586a914ed459dcf970b0" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.445171 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sxb2r"] Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.462924 4987 scope.go:117] "RemoveContainer" containerID="b021bd8051fc2a1ae99fcb55af1e750d94af08fd500642c244e23331d7abbd6c" Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.467331 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sxb2r"] Oct 04 07:34:18 crc kubenswrapper[4987]: I1004 07:34:18.493716 4987 scope.go:117] "RemoveContainer" containerID="a25c25285e05971d23987ac5467f675bf4a5ac5b9f2f873407ed8885edfc6f33" Oct 04 07:34:20 crc kubenswrapper[4987]: I1004 07:34:20.066589 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" path="/var/lib/kubelet/pods/5f539ae6-9dc4-4b26-b7bc-d4052346ec0a/volumes" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.204458 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fhhhq"] Oct 04 07:34:46 crc kubenswrapper[4987]: E1004 07:34:46.205673 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="registry-server" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.205692 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="registry-server" Oct 04 07:34:46 crc kubenswrapper[4987]: E1004 07:34:46.205723 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="extract-utilities" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.205730 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="extract-utilities" Oct 04 07:34:46 crc kubenswrapper[4987]: E1004 07:34:46.205738 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="extract-content" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.205746 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="extract-content" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.205902 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f539ae6-9dc4-4b26-b7bc-d4052346ec0a" containerName="registry-server" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.207094 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.217818 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhhhq"] Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.307658 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-catalog-content\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.308262 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-utilities\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.308300 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6bl4\" (UniqueName: \"kubernetes.io/projected/4396cf17-b80e-4f17-aae5-3043e1c5012b-kube-api-access-q6bl4\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.410304 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-catalog-content\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.410447 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-utilities\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.410476 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6bl4\" (UniqueName: \"kubernetes.io/projected/4396cf17-b80e-4f17-aae5-3043e1c5012b-kube-api-access-q6bl4\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.411216 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-catalog-content\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.411281 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-utilities\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.435846 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6bl4\" (UniqueName: \"kubernetes.io/projected/4396cf17-b80e-4f17-aae5-3043e1c5012b-kube-api-access-q6bl4\") pod \"redhat-marketplace-fhhhq\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.527985 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:46 crc kubenswrapper[4987]: I1004 07:34:46.774493 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhhhq"] Oct 04 07:34:47 crc kubenswrapper[4987]: I1004 07:34:47.404337 4987 scope.go:117] "RemoveContainer" containerID="6b8ad1c4bf6f6f36ac8f306829df4bf04e058c17c03b08d77d58b06236523b6a" Oct 04 07:34:47 crc kubenswrapper[4987]: I1004 07:34:47.706128 4987 generic.go:334] "Generic (PLEG): container finished" podID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerID="110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3" exitCode=0 Oct 04 07:34:47 crc kubenswrapper[4987]: I1004 07:34:47.706205 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhhhq" event={"ID":"4396cf17-b80e-4f17-aae5-3043e1c5012b","Type":"ContainerDied","Data":"110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3"} Oct 04 07:34:47 crc kubenswrapper[4987]: I1004 07:34:47.706386 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhhhq" event={"ID":"4396cf17-b80e-4f17-aae5-3043e1c5012b","Type":"ContainerStarted","Data":"9f9a6f05a2216b5d62719cc464bf9a112663f483d350706de8910944e361d1b5"} Oct 04 07:34:48 crc kubenswrapper[4987]: I1004 07:34:48.718141 4987 generic.go:334] "Generic (PLEG): container finished" podID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerID="fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321" exitCode=0 Oct 04 07:34:48 crc kubenswrapper[4987]: I1004 07:34:48.718258 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhhhq" event={"ID":"4396cf17-b80e-4f17-aae5-3043e1c5012b","Type":"ContainerDied","Data":"fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321"} Oct 04 07:34:49 crc kubenswrapper[4987]: I1004 07:34:49.735823 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhhhq" event={"ID":"4396cf17-b80e-4f17-aae5-3043e1c5012b","Type":"ContainerStarted","Data":"e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90"} Oct 04 07:34:49 crc kubenswrapper[4987]: I1004 07:34:49.767296 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fhhhq" podStartSLOduration=2.262035032 podStartE2EDuration="3.767274398s" podCreationTimestamp="2025-10-04 07:34:46 +0000 UTC" firstStartedPulling="2025-10-04 07:34:47.709311107 +0000 UTC m=+1504.764209959" lastFinishedPulling="2025-10-04 07:34:49.214550463 +0000 UTC m=+1506.269449325" observedRunningTime="2025-10-04 07:34:49.76573898 +0000 UTC m=+1506.820637832" watchObservedRunningTime="2025-10-04 07:34:49.767274398 +0000 UTC m=+1506.822173240" Oct 04 07:34:56 crc kubenswrapper[4987]: I1004 07:34:56.528563 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:56 crc kubenswrapper[4987]: I1004 07:34:56.529463 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:56 crc kubenswrapper[4987]: I1004 07:34:56.590811 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:56 crc kubenswrapper[4987]: I1004 07:34:56.865535 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:56 crc kubenswrapper[4987]: I1004 07:34:56.936012 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhhhq"] Oct 04 07:34:58 crc kubenswrapper[4987]: I1004 07:34:58.823312 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fhhhq" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="registry-server" containerID="cri-o://e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90" gracePeriod=2 Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.256923 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.349439 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-catalog-content\") pod \"4396cf17-b80e-4f17-aae5-3043e1c5012b\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.349507 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-utilities\") pod \"4396cf17-b80e-4f17-aae5-3043e1c5012b\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.349581 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6bl4\" (UniqueName: \"kubernetes.io/projected/4396cf17-b80e-4f17-aae5-3043e1c5012b-kube-api-access-q6bl4\") pod \"4396cf17-b80e-4f17-aae5-3043e1c5012b\" (UID: \"4396cf17-b80e-4f17-aae5-3043e1c5012b\") " Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.350700 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-utilities" (OuterVolumeSpecName: "utilities") pod "4396cf17-b80e-4f17-aae5-3043e1c5012b" (UID: "4396cf17-b80e-4f17-aae5-3043e1c5012b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.360095 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4396cf17-b80e-4f17-aae5-3043e1c5012b-kube-api-access-q6bl4" (OuterVolumeSpecName: "kube-api-access-q6bl4") pod "4396cf17-b80e-4f17-aae5-3043e1c5012b" (UID: "4396cf17-b80e-4f17-aae5-3043e1c5012b"). InnerVolumeSpecName "kube-api-access-q6bl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.372718 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4396cf17-b80e-4f17-aae5-3043e1c5012b" (UID: "4396cf17-b80e-4f17-aae5-3043e1c5012b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.452417 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.452495 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4396cf17-b80e-4f17-aae5-3043e1c5012b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.452529 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6bl4\" (UniqueName: \"kubernetes.io/projected/4396cf17-b80e-4f17-aae5-3043e1c5012b-kube-api-access-q6bl4\") on node \"crc\" DevicePath \"\"" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.836569 4987 generic.go:334] "Generic (PLEG): container finished" podID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerID="e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90" exitCode=0 Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.836650 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhhhq" event={"ID":"4396cf17-b80e-4f17-aae5-3043e1c5012b","Type":"ContainerDied","Data":"e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90"} Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.836724 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fhhhq" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.836761 4987 scope.go:117] "RemoveContainer" containerID="e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.836739 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fhhhq" event={"ID":"4396cf17-b80e-4f17-aae5-3043e1c5012b","Type":"ContainerDied","Data":"9f9a6f05a2216b5d62719cc464bf9a112663f483d350706de8910944e361d1b5"} Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.877229 4987 scope.go:117] "RemoveContainer" containerID="fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.883488 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhhhq"] Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.890097 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fhhhq"] Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.909761 4987 scope.go:117] "RemoveContainer" containerID="110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.950744 4987 scope.go:117] "RemoveContainer" containerID="e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90" Oct 04 07:34:59 crc kubenswrapper[4987]: E1004 07:34:59.951271 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90\": container with ID starting with e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90 not found: ID does not exist" containerID="e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.951310 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90"} err="failed to get container status \"e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90\": rpc error: code = NotFound desc = could not find container \"e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90\": container with ID starting with e2c80b109b0d64b9a7cabaeb1546362c70a00bc45ebd70047c218433bb818d90 not found: ID does not exist" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.951339 4987 scope.go:117] "RemoveContainer" containerID="fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321" Oct 04 07:34:59 crc kubenswrapper[4987]: E1004 07:34:59.951777 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321\": container with ID starting with fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321 not found: ID does not exist" containerID="fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.951843 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321"} err="failed to get container status \"fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321\": rpc error: code = NotFound desc = could not find container \"fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321\": container with ID starting with fda686aca18056bc15b88ebe363bbb2634d6335bc44c9e7929bb42b16c440321 not found: ID does not exist" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.951915 4987 scope.go:117] "RemoveContainer" containerID="110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3" Oct 04 07:34:59 crc kubenswrapper[4987]: E1004 07:34:59.952328 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3\": container with ID starting with 110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3 not found: ID does not exist" containerID="110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3" Oct 04 07:34:59 crc kubenswrapper[4987]: I1004 07:34:59.952368 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3"} err="failed to get container status \"110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3\": rpc error: code = NotFound desc = could not find container \"110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3\": container with ID starting with 110cf1693a0cd7c8ecd185f3eaabeba3b591ebb78f2d8f418e19c8dd03bdf9e3 not found: ID does not exist" Oct 04 07:35:00 crc kubenswrapper[4987]: I1004 07:35:00.068007 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" path="/var/lib/kubelet/pods/4396cf17-b80e-4f17-aae5-3043e1c5012b/volumes" Oct 04 07:35:12 crc kubenswrapper[4987]: I1004 07:35:12.703742 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:35:12 crc kubenswrapper[4987]: I1004 07:35:12.704460 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:35:42 crc kubenswrapper[4987]: I1004 07:35:42.703301 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:35:42 crc kubenswrapper[4987]: I1004 07:35:42.704175 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:35:47 crc kubenswrapper[4987]: I1004 07:35:47.483986 4987 scope.go:117] "RemoveContainer" containerID="a188687e0e4e09e1ccc47d4d9c2d7d26742ec18517d2ccc0286fe72ec878db10" Oct 04 07:35:47 crc kubenswrapper[4987]: I1004 07:35:47.516590 4987 scope.go:117] "RemoveContainer" containerID="fc10285d1d547f376c0cf0fce7788cb81b08854e00c7b5cbe8869d391fe4ee71" Oct 04 07:36:12 crc kubenswrapper[4987]: I1004 07:36:12.703085 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:36:12 crc kubenswrapper[4987]: I1004 07:36:12.703700 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:36:12 crc kubenswrapper[4987]: I1004 07:36:12.703761 4987 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" Oct 04 07:36:12 crc kubenswrapper[4987]: I1004 07:36:12.704442 4987 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d"} pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 07:36:12 crc kubenswrapper[4987]: I1004 07:36:12.704496 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" containerID="cri-o://c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" gracePeriod=600 Oct 04 07:36:12 crc kubenswrapper[4987]: E1004 07:36:12.828742 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:36:13 crc kubenswrapper[4987]: I1004 07:36:13.552330 4987 generic.go:334] "Generic (PLEG): container finished" podID="a89a1560-c018-4df7-92aa-795638d35d94" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" exitCode=0 Oct 04 07:36:13 crc kubenswrapper[4987]: I1004 07:36:13.552452 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerDied","Data":"c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d"} Oct 04 07:36:13 crc kubenswrapper[4987]: I1004 07:36:13.552786 4987 scope.go:117] "RemoveContainer" containerID="901e84e5b40d7f1df5e525731897cdcc9e887088810c1590e56d6e773c746b71" Oct 04 07:36:13 crc kubenswrapper[4987]: I1004 07:36:13.556185 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:36:13 crc kubenswrapper[4987]: E1004 07:36:13.557701 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:36:26 crc kubenswrapper[4987]: I1004 07:36:26.054699 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:36:26 crc kubenswrapper[4987]: E1004 07:36:26.055979 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:36:38 crc kubenswrapper[4987]: I1004 07:36:38.053607 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:36:38 crc kubenswrapper[4987]: E1004 07:36:38.054778 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:36:46 crc kubenswrapper[4987]: I1004 07:36:46.069505 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-cww6m"] Oct 04 07:36:46 crc kubenswrapper[4987]: I1004 07:36:46.075609 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-cww6m"] Oct 04 07:36:47 crc kubenswrapper[4987]: I1004 07:36:47.637008 4987 scope.go:117] "RemoveContainer" containerID="04fdc735ad6ebffb6c35ed6aa6c6333b4e7f90a581a8fe4baec5f6aeeb3a50c3" Oct 04 07:36:47 crc kubenswrapper[4987]: I1004 07:36:47.672456 4987 scope.go:117] "RemoveContainer" containerID="230f7148684aca1d5c37104a266592388ff589f87dcb05b50527544b44898f6f" Oct 04 07:36:47 crc kubenswrapper[4987]: I1004 07:36:47.702580 4987 scope.go:117] "RemoveContainer" containerID="420ec6669ee583256e6cd9c5022954413875625e228f48012862a95cbadd77ff" Oct 04 07:36:47 crc kubenswrapper[4987]: I1004 07:36:47.785815 4987 scope.go:117] "RemoveContainer" containerID="64077221f913273b05f07b08432f831cece5d5d0e65bd7ebc1ccab4052a008cb" Oct 04 07:36:47 crc kubenswrapper[4987]: I1004 07:36:47.849674 4987 scope.go:117] "RemoveContainer" containerID="65d310628aac2f405df3ebf65965e0164c5df4a60d34ad4c46eef77bba52ab06" Oct 04 07:36:48 crc kubenswrapper[4987]: I1004 07:36:48.062716 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a93fd53a-dad1-4a5a-8e9e-50c393cac9b2" path="/var/lib/kubelet/pods/a93fd53a-dad1-4a5a-8e9e-50c393cac9b2/volumes" Oct 04 07:36:53 crc kubenswrapper[4987]: I1004 07:36:53.054923 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:36:53 crc kubenswrapper[4987]: E1004 07:36:53.055511 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:36:55 crc kubenswrapper[4987]: I1004 07:36:55.033691 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-c0c2-account-create-blbdj"] Oct 04 07:36:55 crc kubenswrapper[4987]: I1004 07:36:55.040136 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-c0c2-account-create-blbdj"] Oct 04 07:36:56 crc kubenswrapper[4987]: I1004 07:36:56.064550 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eff36d2-e701-4996-b283-cfbbc1b27e0c" path="/var/lib/kubelet/pods/9eff36d2-e701-4996-b283-cfbbc1b27e0c/volumes" Oct 04 07:37:05 crc kubenswrapper[4987]: I1004 07:37:05.053969 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:37:05 crc kubenswrapper[4987]: E1004 07:37:05.054961 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:37:18 crc kubenswrapper[4987]: I1004 07:37:18.041327 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-6sb9w"] Oct 04 07:37:18 crc kubenswrapper[4987]: I1004 07:37:18.047302 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-6sb9w"] Oct 04 07:37:18 crc kubenswrapper[4987]: I1004 07:37:18.064120 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91962d41-2412-432a-823e-a4351de82e52" path="/var/lib/kubelet/pods/91962d41-2412-432a-823e-a4351de82e52/volumes" Oct 04 07:37:19 crc kubenswrapper[4987]: I1004 07:37:19.054407 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:37:19 crc kubenswrapper[4987]: E1004 07:37:19.054737 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:37:26 crc kubenswrapper[4987]: I1004 07:37:26.026991 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-c27lb"] Oct 04 07:37:26 crc kubenswrapper[4987]: I1004 07:37:26.031374 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-c27lb"] Oct 04 07:37:26 crc kubenswrapper[4987]: I1004 07:37:26.065141 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="814ffbb2-3f6a-4658-9e65-9c592d4d5b92" path="/var/lib/kubelet/pods/814ffbb2-3f6a-4658-9e65-9c592d4d5b92/volumes" Oct 04 07:37:33 crc kubenswrapper[4987]: I1004 07:37:33.053664 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:37:33 crc kubenswrapper[4987]: E1004 07:37:33.056436 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:37:45 crc kubenswrapper[4987]: I1004 07:37:45.054481 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:37:45 crc kubenswrapper[4987]: E1004 07:37:45.056031 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:37:48 crc kubenswrapper[4987]: I1004 07:37:48.019003 4987 scope.go:117] "RemoveContainer" containerID="6ae61230aa2866af4eb23996387d0c3b70dc594f1a9694ccb5009cfc3fe0abf8" Oct 04 07:37:48 crc kubenswrapper[4987]: I1004 07:37:48.055582 4987 scope.go:117] "RemoveContainer" containerID="f50f5cdd0a355c36be1e6022a1de82292de621a28d536a0f7bb0735e03d1e5be" Oct 04 07:37:48 crc kubenswrapper[4987]: I1004 07:37:48.097268 4987 scope.go:117] "RemoveContainer" containerID="d90ead49c8426aa3aee98ecb42a5e8a0b92d1c9f822ec4e481ab244c9ee1aead" Oct 04 07:37:48 crc kubenswrapper[4987]: I1004 07:37:48.139258 4987 scope.go:117] "RemoveContainer" containerID="b380672c2863904357879ef19930b152e78463e79fe93dfe15304bdd7fc36157" Oct 04 07:37:48 crc kubenswrapper[4987]: I1004 07:37:48.160107 4987 scope.go:117] "RemoveContainer" containerID="14a1def2b6c691fa8461c6ab4d716c08f3b62a75097feef3c4dd69676fee4592" Oct 04 07:37:56 crc kubenswrapper[4987]: I1004 07:37:56.053761 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:37:56 crc kubenswrapper[4987]: E1004 07:37:56.054760 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.039144 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:38:02 crc kubenswrapper[4987]: E1004 07:38:02.040168 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="registry-server" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.040188 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="registry-server" Oct 04 07:38:02 crc kubenswrapper[4987]: E1004 07:38:02.040227 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="extract-utilities" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.040234 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="extract-utilities" Oct 04 07:38:02 crc kubenswrapper[4987]: E1004 07:38:02.040248 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="extract-content" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.040255 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="extract-content" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.040396 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="4396cf17-b80e-4f17-aae5-3043e1c5012b" containerName="registry-server" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.041076 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.046117 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.046855 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.046976 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-sfbrn" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.046863 4987 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.051308 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.201980 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-config-secret\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.202097 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-config\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.202160 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jvhx\" (UniqueName: \"kubernetes.io/projected/ad686ae4-6169-4283-afb3-664cd306fb4d-kube-api-access-2jvhx\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.202349 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-scripts\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.304293 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jvhx\" (UniqueName: \"kubernetes.io/projected/ad686ae4-6169-4283-afb3-664cd306fb4d-kube-api-access-2jvhx\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.304405 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-scripts\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.304441 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-config-secret\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.304482 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-config\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.305502 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-config\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.306487 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-scripts\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.310237 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ad686ae4-6169-4283-afb3-664cd306fb4d-openstack-config-secret\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.320971 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jvhx\" (UniqueName: \"kubernetes.io/projected/ad686ae4-6169-4283-afb3-664cd306fb4d-kube-api-access-2jvhx\") pod \"openstackclient\" (UID: \"ad686ae4-6169-4283-afb3-664cd306fb4d\") " pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.366437 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Oct 04 07:38:02 crc kubenswrapper[4987]: I1004 07:38:02.590359 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Oct 04 07:38:03 crc kubenswrapper[4987]: I1004 07:38:03.588260 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"ad686ae4-6169-4283-afb3-664cd306fb4d","Type":"ContainerStarted","Data":"01e8d4ba729e8805c1c5d0ddfc1496d4d9d97802da31e725ea9cda0bd4f5e883"} Oct 04 07:38:03 crc kubenswrapper[4987]: I1004 07:38:03.588673 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"ad686ae4-6169-4283-afb3-664cd306fb4d","Type":"ContainerStarted","Data":"250d367c364bddb9e4cf45b748cf28109fba2f7d23cd20ba7b7d6a99e37991e4"} Oct 04 07:38:03 crc kubenswrapper[4987]: I1004 07:38:03.615499 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.615466021 podStartE2EDuration="1.615466021s" podCreationTimestamp="2025-10-04 07:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 07:38:03.603459931 +0000 UTC m=+1700.658358813" watchObservedRunningTime="2025-10-04 07:38:03.615466021 +0000 UTC m=+1700.670364893" Oct 04 07:38:07 crc kubenswrapper[4987]: I1004 07:38:07.054672 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:38:07 crc kubenswrapper[4987]: E1004 07:38:07.055318 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:38:20 crc kubenswrapper[4987]: I1004 07:38:20.054266 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:38:20 crc kubenswrapper[4987]: E1004 07:38:20.055172 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:38:34 crc kubenswrapper[4987]: I1004 07:38:34.054111 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:38:34 crc kubenswrapper[4987]: E1004 07:38:34.054969 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:38:46 crc kubenswrapper[4987]: I1004 07:38:46.054229 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:38:46 crc kubenswrapper[4987]: E1004 07:38:46.055120 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:39:00 crc kubenswrapper[4987]: I1004 07:39:00.053672 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:39:00 crc kubenswrapper[4987]: E1004 07:39:00.054431 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.230990 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tdcvz"] Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.235467 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.251508 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdcvz"] Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.286618 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-catalog-content\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.286807 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-utilities\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.286878 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t574j\" (UniqueName: \"kubernetes.io/projected/c1b2977f-d89a-444b-89ed-7651d99b20a7-kube-api-access-t574j\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.389002 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-catalog-content\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.389059 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-utilities\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.389111 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t574j\" (UniqueName: \"kubernetes.io/projected/c1b2977f-d89a-444b-89ed-7651d99b20a7-kube-api-access-t574j\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.389600 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-catalog-content\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.389648 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-utilities\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.412578 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t574j\" (UniqueName: \"kubernetes.io/projected/c1b2977f-d89a-444b-89ed-7651d99b20a7-kube-api-access-t574j\") pod \"certified-operators-tdcvz\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.558825 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:06 crc kubenswrapper[4987]: I1004 07:39:06.907094 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdcvz"] Oct 04 07:39:07 crc kubenswrapper[4987]: I1004 07:39:07.120013 4987 generic.go:334] "Generic (PLEG): container finished" podID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerID="c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb" exitCode=0 Oct 04 07:39:07 crc kubenswrapper[4987]: I1004 07:39:07.120098 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdcvz" event={"ID":"c1b2977f-d89a-444b-89ed-7651d99b20a7","Type":"ContainerDied","Data":"c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb"} Oct 04 07:39:07 crc kubenswrapper[4987]: I1004 07:39:07.120537 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdcvz" event={"ID":"c1b2977f-d89a-444b-89ed-7651d99b20a7","Type":"ContainerStarted","Data":"71fe7e32d0dae973f7fb4ecb3039040bbeee82856d9680af7385585d761ed96f"} Oct 04 07:39:07 crc kubenswrapper[4987]: I1004 07:39:07.124547 4987 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 07:39:09 crc kubenswrapper[4987]: I1004 07:39:09.138012 4987 generic.go:334] "Generic (PLEG): container finished" podID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerID="3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e" exitCode=0 Oct 04 07:39:09 crc kubenswrapper[4987]: I1004 07:39:09.138140 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdcvz" event={"ID":"c1b2977f-d89a-444b-89ed-7651d99b20a7","Type":"ContainerDied","Data":"3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e"} Oct 04 07:39:11 crc kubenswrapper[4987]: I1004 07:39:11.156823 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdcvz" event={"ID":"c1b2977f-d89a-444b-89ed-7651d99b20a7","Type":"ContainerStarted","Data":"3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6"} Oct 04 07:39:11 crc kubenswrapper[4987]: I1004 07:39:11.182429 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tdcvz" podStartSLOduration=2.106784146 podStartE2EDuration="5.182398852s" podCreationTimestamp="2025-10-04 07:39:06 +0000 UTC" firstStartedPulling="2025-10-04 07:39:07.123549671 +0000 UTC m=+1764.178448503" lastFinishedPulling="2025-10-04 07:39:10.199164347 +0000 UTC m=+1767.254063209" observedRunningTime="2025-10-04 07:39:11.177677472 +0000 UTC m=+1768.232576334" watchObservedRunningTime="2025-10-04 07:39:11.182398852 +0000 UTC m=+1768.237297704" Oct 04 07:39:13 crc kubenswrapper[4987]: I1004 07:39:13.053583 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:39:13 crc kubenswrapper[4987]: E1004 07:39:13.054222 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:39:16 crc kubenswrapper[4987]: I1004 07:39:16.559465 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:16 crc kubenswrapper[4987]: I1004 07:39:16.560081 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:16 crc kubenswrapper[4987]: I1004 07:39:16.622515 4987 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:17 crc kubenswrapper[4987]: I1004 07:39:17.259362 4987 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:17 crc kubenswrapper[4987]: I1004 07:39:17.311699 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tdcvz"] Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.219706 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tdcvz" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="registry-server" containerID="cri-o://3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6" gracePeriod=2 Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.650389 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.832562 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-utilities\") pod \"c1b2977f-d89a-444b-89ed-7651d99b20a7\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.832740 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t574j\" (UniqueName: \"kubernetes.io/projected/c1b2977f-d89a-444b-89ed-7651d99b20a7-kube-api-access-t574j\") pod \"c1b2977f-d89a-444b-89ed-7651d99b20a7\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.832771 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-catalog-content\") pod \"c1b2977f-d89a-444b-89ed-7651d99b20a7\" (UID: \"c1b2977f-d89a-444b-89ed-7651d99b20a7\") " Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.834217 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-utilities" (OuterVolumeSpecName: "utilities") pod "c1b2977f-d89a-444b-89ed-7651d99b20a7" (UID: "c1b2977f-d89a-444b-89ed-7651d99b20a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.855943 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b2977f-d89a-444b-89ed-7651d99b20a7-kube-api-access-t574j" (OuterVolumeSpecName: "kube-api-access-t574j") pod "c1b2977f-d89a-444b-89ed-7651d99b20a7" (UID: "c1b2977f-d89a-444b-89ed-7651d99b20a7"). InnerVolumeSpecName "kube-api-access-t574j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.934902 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t574j\" (UniqueName: \"kubernetes.io/projected/c1b2977f-d89a-444b-89ed-7651d99b20a7-kube-api-access-t574j\") on node \"crc\" DevicePath \"\"" Oct 04 07:39:19 crc kubenswrapper[4987]: I1004 07:39:19.936071 4987 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.233822 4987 generic.go:334] "Generic (PLEG): container finished" podID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerID="3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6" exitCode=0 Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.233889 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdcvz" event={"ID":"c1b2977f-d89a-444b-89ed-7651d99b20a7","Type":"ContainerDied","Data":"3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6"} Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.233948 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdcvz" event={"ID":"c1b2977f-d89a-444b-89ed-7651d99b20a7","Type":"ContainerDied","Data":"71fe7e32d0dae973f7fb4ecb3039040bbeee82856d9680af7385585d761ed96f"} Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.234013 4987 scope.go:117] "RemoveContainer" containerID="3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.234212 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdcvz" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.273970 4987 scope.go:117] "RemoveContainer" containerID="3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.301536 4987 scope.go:117] "RemoveContainer" containerID="c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.345712 4987 scope.go:117] "RemoveContainer" containerID="3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6" Oct 04 07:39:20 crc kubenswrapper[4987]: E1004 07:39:20.346319 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6\": container with ID starting with 3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6 not found: ID does not exist" containerID="3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.346360 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6"} err="failed to get container status \"3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6\": rpc error: code = NotFound desc = could not find container \"3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6\": container with ID starting with 3e84bcb86401398d934776ba0caefa79bd52a19807706f5796c5c5439f1c5ba6 not found: ID does not exist" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.346393 4987 scope.go:117] "RemoveContainer" containerID="3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e" Oct 04 07:39:20 crc kubenswrapper[4987]: E1004 07:39:20.346700 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e\": container with ID starting with 3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e not found: ID does not exist" containerID="3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.346757 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e"} err="failed to get container status \"3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e\": rpc error: code = NotFound desc = could not find container \"3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e\": container with ID starting with 3bdb6a8f361f71998e7a257689f7ab542af75188200e13c79dd86e901b2e350e not found: ID does not exist" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.346799 4987 scope.go:117] "RemoveContainer" containerID="c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb" Oct 04 07:39:20 crc kubenswrapper[4987]: E1004 07:39:20.347813 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb\": container with ID starting with c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb not found: ID does not exist" containerID="c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.347843 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb"} err="failed to get container status \"c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb\": rpc error: code = NotFound desc = could not find container \"c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb\": container with ID starting with c9f0148ae9d9253585b1ac8a9b50deba34e0ac5added917c777608d338b93bcb not found: ID does not exist" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.499591 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1b2977f-d89a-444b-89ed-7651d99b20a7" (UID: "c1b2977f-d89a-444b-89ed-7651d99b20a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.547164 4987 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1b2977f-d89a-444b-89ed-7651d99b20a7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.566121 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tdcvz"] Oct 04 07:39:20 crc kubenswrapper[4987]: I1004 07:39:20.573747 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tdcvz"] Oct 04 07:39:22 crc kubenswrapper[4987]: I1004 07:39:22.063059 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" path="/var/lib/kubelet/pods/c1b2977f-d89a-444b-89ed-7651d99b20a7/volumes" Oct 04 07:39:27 crc kubenswrapper[4987]: I1004 07:39:27.053696 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:39:27 crc kubenswrapper[4987]: E1004 07:39:27.054480 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:39:38 crc kubenswrapper[4987]: I1004 07:39:38.054002 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:39:38 crc kubenswrapper[4987]: E1004 07:39:38.055288 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.096575 4987 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gvjbx/must-gather-nc5f2"] Oct 04 07:39:39 crc kubenswrapper[4987]: E1004 07:39:39.097490 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="extract-content" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.097536 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="extract-content" Oct 04 07:39:39 crc kubenswrapper[4987]: E1004 07:39:39.097562 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="registry-server" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.097577 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="registry-server" Oct 04 07:39:39 crc kubenswrapper[4987]: E1004 07:39:39.097603 4987 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="extract-utilities" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.097617 4987 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="extract-utilities" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.101669 4987 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b2977f-d89a-444b-89ed-7651d99b20a7" containerName="registry-server" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.102936 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.110594 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gvjbx"/"openshift-service-ca.crt" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.120506 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gvjbx/must-gather-nc5f2"] Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.122377 4987 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gvjbx"/"kube-root-ca.crt" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.204874 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9hw2\" (UniqueName: \"kubernetes.io/projected/b67cc4cf-a982-4378-81cf-e179d0149f84-kube-api-access-r9hw2\") pod \"must-gather-nc5f2\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.204946 4987 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b67cc4cf-a982-4378-81cf-e179d0149f84-must-gather-output\") pod \"must-gather-nc5f2\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.308564 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9hw2\" (UniqueName: \"kubernetes.io/projected/b67cc4cf-a982-4378-81cf-e179d0149f84-kube-api-access-r9hw2\") pod \"must-gather-nc5f2\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.308679 4987 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b67cc4cf-a982-4378-81cf-e179d0149f84-must-gather-output\") pod \"must-gather-nc5f2\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.309324 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b67cc4cf-a982-4378-81cf-e179d0149f84-must-gather-output\") pod \"must-gather-nc5f2\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.331260 4987 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9hw2\" (UniqueName: \"kubernetes.io/projected/b67cc4cf-a982-4378-81cf-e179d0149f84-kube-api-access-r9hw2\") pod \"must-gather-nc5f2\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.428967 4987 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:39:39 crc kubenswrapper[4987]: I1004 07:39:39.672716 4987 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gvjbx/must-gather-nc5f2"] Oct 04 07:39:40 crc kubenswrapper[4987]: I1004 07:39:40.435516 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" event={"ID":"b67cc4cf-a982-4378-81cf-e179d0149f84","Type":"ContainerStarted","Data":"2c8e4d126a9c3d6623166afc10644e2675ec24fa3058d9145e5c828e893765df"} Oct 04 07:39:44 crc kubenswrapper[4987]: I1004 07:39:44.469017 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" event={"ID":"b67cc4cf-a982-4378-81cf-e179d0149f84","Type":"ContainerStarted","Data":"6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4"} Oct 04 07:39:44 crc kubenswrapper[4987]: I1004 07:39:44.469695 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" event={"ID":"b67cc4cf-a982-4378-81cf-e179d0149f84","Type":"ContainerStarted","Data":"0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3"} Oct 04 07:39:44 crc kubenswrapper[4987]: I1004 07:39:44.486019 4987 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" podStartSLOduration=1.716517097 podStartE2EDuration="5.485991012s" podCreationTimestamp="2025-10-04 07:39:39 +0000 UTC" firstStartedPulling="2025-10-04 07:39:39.704284685 +0000 UTC m=+1796.759183527" lastFinishedPulling="2025-10-04 07:39:43.4737586 +0000 UTC m=+1800.528657442" observedRunningTime="2025-10-04 07:39:44.484176575 +0000 UTC m=+1801.539075417" watchObservedRunningTime="2025-10-04 07:39:44.485991012 +0000 UTC m=+1801.540889854" Oct 04 07:39:49 crc kubenswrapper[4987]: I1004 07:39:49.053372 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:39:49 crc kubenswrapper[4987]: E1004 07:39:49.054319 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:40:03 crc kubenswrapper[4987]: I1004 07:40:03.052878 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:40:03 crc kubenswrapper[4987]: E1004 07:40:03.053969 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:40:16 crc kubenswrapper[4987]: I1004 07:40:16.055639 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:40:16 crc kubenswrapper[4987]: E1004 07:40:16.056712 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.247466 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/util/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.358815 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/util/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.407901 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/pull/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.490561 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/pull/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.686895 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/util/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.700413 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/extract/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.710011 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_39c5fb04f82fea7d774420ed8f65200e06c90bf50d6a3f049bcdcccd8f7rqjq_fd50d7a6-1e67-4764-b700-39081ed2da4a/pull/0.log" Oct 04 07:40:20 crc kubenswrapper[4987]: I1004 07:40:20.888973 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/util/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.077262 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/pull/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.095602 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/pull/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.109198 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/util/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.252530 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/util/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.297499 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/pull/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.307163 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5bfb89b7a15e902ec1ce651098a1cbdcb0a2281c38e30d9a342b952813jfjv9_31d9f706-030b-4264-bf7f-7847c676a095/extract/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.441911 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/util/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.644101 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/util/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.647652 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/pull/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.665575 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/pull/0.log" Oct 04 07:40:21 crc kubenswrapper[4987]: I1004 07:40:21.994209 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/extract/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.022055 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/util/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.068767 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8da7e2360b27a0f547620d000b3ce05fecd69e104ed121c8b883e332937tdsk_32ec6fc4-2f39-4e20-b529-9aa82a827fb7/pull/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.233196 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/util/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.402994 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/pull/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.432572 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/util/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.469925 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/pull/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.653722 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/util/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.685877 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/pull/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.697538 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590r6pg2_82cd04d6-e9cc-4f3a-89f6-d867a256a4af/extract/0.log" Oct 04 07:40:22 crc kubenswrapper[4987]: I1004 07:40:22.838243 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.011658 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.027971 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.044300 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.223545 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.224704 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/extract/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.284498 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_be478ded7503179b705a514b949fce362db409745a46d496f1170a512cfc6kw_a5d0cde7-7997-4601-b1e1-f8d55d28f8b3/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.357717 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.486234 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.493017 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.526067 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.730460 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/extract/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.743217 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.767160 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.775241 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c77910536a79801a83f49d4fd4581e5a2972791dfc31ed0ea9f0ffea32pldhd_70c28efa-16a5-4eef-8fec-f80bad6cea42/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.969334 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/util/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.970780 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/pull/0.log" Oct 04 07:40:23 crc kubenswrapper[4987]: I1004 07:40:23.989511 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/pull/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.147567 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/util/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.223233 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/extract/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.236980 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6468bd8bb6-p22sc_4f9c4992-f741-41e4-9f30-90c4d1c165a5/kube-rbac-proxy/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.254675 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ec20a04ef7278338c96ca90950ec47944973b8553e1da5c6f2ce730402sqpqp_d19926e0-f3c3-40e3-96e1-9fac953bdea3/pull/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.427174 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-f5msg_21c12a08-bede-4878-b1ad-3f847091bf9a/registry-server/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.464249 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-6468bd8bb6-p22sc_4f9c4992-f741-41e4-9f30-90c4d1c165a5/manager/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.498071 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-687c78db9b-kfz5x_a0639060-01de-41cc-90b7-e1cde85170fc/kube-rbac-proxy/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.642910 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-687c78db9b-kfz5x_a0639060-01de-41cc-90b7-e1cde85170fc/manager/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.681728 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-2kdfj_09609231-954a-4b7a-9878-e4f618347a47/registry-server/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.758944 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-759bcdfbb8-g4mlv_1c819f4f-118f-4592-8f51-b2f564437526/kube-rbac-proxy/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.889851 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-759bcdfbb8-g4mlv_1c819f4f-118f-4592-8f51-b2f564437526/manager/0.log" Oct 04 07:40:24 crc kubenswrapper[4987]: I1004 07:40:24.978561 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-mdkmg_e1582f26-057a-431d-8fff-4e8b0f139149/registry-server/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.063689 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55fd955c48-qdph7_81bd06d1-6dc2-4132-9e01-2d3b72a69cea/kube-rbac-proxy/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.160695 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55fd955c48-qdph7_81bd06d1-6dc2-4132-9e01-2d3b72a69cea/manager/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.222740 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-bcsxf_1d66401a-3587-4fdd-9a4c-fac02692709d/registry-server/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.315315 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b74947f6c-qsx2n_bb4722ac-25c0-48c9-86aa-35576a0aa6e5/kube-rbac-proxy/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.384995 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b74947f6c-qsx2n_bb4722ac-25c0-48c9-86aa-35576a0aa6e5/manager/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.461660 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-n2x7n_092dd76b-510c-49fb-be46-5a8ebfb818a7/registry-server/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.528165 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-rvmfx_d3f2f50a-bd5f-4182-af2f-edd6bce09875/operator/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.633310 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-j9wbv_3dc91b62-1cd4-40db-b853-ff83dd36cd0d/registry-server/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.757217 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6dcf9c964f-9rbnc_d3c896ee-c16b-47c4-b74a-706168509fb5/kube-rbac-proxy/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.781381 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6dcf9c964f-9rbnc_d3c896ee-c16b-47c4-b74a-706168509fb5/manager/0.log" Oct 04 07:40:25 crc kubenswrapper[4987]: I1004 07:40:25.887044 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-qjl2t_73a59a02-da82-4edc-bded-8d2fc1484bfd/registry-server/0.log" Oct 04 07:40:31 crc kubenswrapper[4987]: I1004 07:40:31.053293 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:40:31 crc kubenswrapper[4987]: E1004 07:40:31.054435 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:40:41 crc kubenswrapper[4987]: I1004 07:40:41.233144 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gdbxg_5420a718-1d0a-4b72-b57d-a8a4a18e7763/control-plane-machine-set-operator/0.log" Oct 04 07:40:41 crc kubenswrapper[4987]: I1004 07:40:41.655095 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f49kw_d5787789-e288-4011-baeb-17da6439f196/kube-rbac-proxy/0.log" Oct 04 07:40:41 crc kubenswrapper[4987]: I1004 07:40:41.697244 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f49kw_d5787789-e288-4011-baeb-17da6439f196/machine-api-operator/0.log" Oct 04 07:40:44 crc kubenswrapper[4987]: I1004 07:40:44.058847 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:40:44 crc kubenswrapper[4987]: E1004 07:40:44.059653 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:40:57 crc kubenswrapper[4987]: I1004 07:40:57.959571 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kx8k7_ebd2db86-e065-412a-aa7b-dd13f4359b1f/kube-rbac-proxy/0.log" Oct 04 07:40:57 crc kubenswrapper[4987]: I1004 07:40:57.986556 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-kx8k7_ebd2db86-e065-412a-aa7b-dd13f4359b1f/controller/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.186072 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-frr-files/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.434085 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-reloader/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.439192 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-metrics/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.491169 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-frr-files/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.527698 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-reloader/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.640178 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-frr-files/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.657105 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-reloader/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.711072 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-metrics/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.718203 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-metrics/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.898861 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-reloader/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.899918 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-metrics/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.932556 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/cp-frr-files/0.log" Oct 04 07:40:58 crc kubenswrapper[4987]: I1004 07:40:58.941732 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/controller/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.053668 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:40:59 crc kubenswrapper[4987]: E1004 07:40:59.054204 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.074305 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/frr-metrics/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.270214 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/kube-rbac-proxy-frr/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.287054 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/kube-rbac-proxy/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.429307 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/reloader/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.509475 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-6lqfh_e4a5ec9a-ead6-49d2-a31b-e28597d4c038/frr-k8s-webhook-server/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.588806 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-fdd4g_e336de0c-93ab-4084-8d1a-1afcc48d7c0e/frr/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.690264 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7f485cfd66-xwgrt_dfbc115b-ff8f-424d-9f9d-d5e21ed2a790/manager/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.800713 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6979fbb46b-pxmhb_1222617f-db7d-4014-b952-79765d041405/webhook-server/0.log" Oct 04 07:40:59 crc kubenswrapper[4987]: I1004 07:40:59.904299 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qj8gd_47b35947-1821-4bf6-93cc-11c49e257911/kube-rbac-proxy/0.log" Oct 04 07:41:00 crc kubenswrapper[4987]: I1004 07:41:00.020545 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qj8gd_47b35947-1821-4bf6-93cc-11c49e257911/speaker/0.log" Oct 04 07:41:11 crc kubenswrapper[4987]: I1004 07:41:11.057421 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:41:11 crc kubenswrapper[4987]: E1004 07:41:11.060068 4987 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-dvv6j_openshift-machine-config-operator(a89a1560-c018-4df7-92aa-795638d35d94)\"" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" Oct 04 07:41:13 crc kubenswrapper[4987]: I1004 07:41:13.675077 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-9e77-account-create-656md_cc2559bd-1713-4220-b9d2-41259f9b5d54/mariadb-account-create/0.log" Oct 04 07:41:13 crc kubenswrapper[4987]: I1004 07:41:13.792749 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-llv7g_47de6379-1946-4012-afab-f15535d50880/mariadb-database-create/0.log" Oct 04 07:41:13 crc kubenswrapper[4987]: I1004 07:41:13.890152 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-q4svf_d30bee49-7223-4652-9615-7155381ad95b/glance-db-sync/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.016965 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26/glance-api/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.049476 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26/glance-httpd/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.111996 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_6d400a6b-b3eb-4cfb-bb07-5ad4e6522a26/glance-log/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.214707 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_d6b0984a-9602-4187-b515-f862cf9f2b4b/glance-api/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.269274 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_d6b0984a-9602-4187-b515-f862cf9f2b4b/glance-httpd/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.365104 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_d6b0984a-9602-4187-b515-f862cf9f2b4b/glance-log/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.661531 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-6fccb96676-gqppm_4851c21f-3f97-4813-b27f-64d7e7e2b764/keystone-api/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.732142 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_d20ac66d-0466-4478-af96-72fc6adc11d8/memcached/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.808473 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_d7b90541-0826-4c5c-a22a-ebebb127e444/mysql-bootstrap/0.log" Oct 04 07:41:14 crc kubenswrapper[4987]: I1004 07:41:14.994838 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_d7b90541-0826-4c5c-a22a-ebebb127e444/galera/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.035440 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_d7b90541-0826-4c5c-a22a-ebebb127e444/mysql-bootstrap/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.080999 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_d6684ada-c0e3-4f8f-a804-80f711240730/mysql-bootstrap/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.442153 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_d6684ada-c0e3-4f8f-a804-80f711240730/mysql-bootstrap/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.467594 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_d6684ada-c0e3-4f8f-a804-80f711240730/galera/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.498509 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_d5ddfd58-951f-46d4-804b-0ae9998befc1/mysql-bootstrap/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.699135 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_d5ddfd58-951f-46d4-804b-0ae9998befc1/mysql-bootstrap/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.750748 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_d5ddfd58-951f-46d4-804b-0ae9998befc1/galera/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.795947 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_ad686ae4-6169-4283-afb3-664cd306fb4d/openstackclient/0.log" Oct 04 07:41:15 crc kubenswrapper[4987]: I1004 07:41:15.954765 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_bcde8202-83ce-4dfa-9562-4733938935f0/setup-container/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.103470 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_bcde8202-83ce-4dfa-9562-4733938935f0/setup-container/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.148610 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_bcde8202-83ce-4dfa-9562-4733938935f0/rabbitmq/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.225705 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-59cb459c9f-qhgk9_a7c67d2f-aa98-45e4-8dc4-ba02f6051378/proxy-httpd/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.293529 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-59cb459c9f-qhgk9_a7c67d2f-aa98-45e4-8dc4-ba02f6051378/proxy-server/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.328555 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-zj4nq_a531f5c9-a54b-4016-83d4-57e4ced832d6/swift-ring-rebalance/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.494685 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/account-auditor/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.527740 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/account-reaper/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.636276 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/account-replicator/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.702202 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/account-server/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.712610 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/container-auditor/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.751788 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/container-replicator/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.846670 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/container-server/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.897937 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/container-updater/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.912432 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/object-auditor/0.log" Oct 04 07:41:16 crc kubenswrapper[4987]: I1004 07:41:16.997995 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/object-expirer/0.log" Oct 04 07:41:17 crc kubenswrapper[4987]: I1004 07:41:17.070036 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/object-replicator/0.log" Oct 04 07:41:17 crc kubenswrapper[4987]: I1004 07:41:17.088392 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/object-server/0.log" Oct 04 07:41:17 crc kubenswrapper[4987]: I1004 07:41:17.093140 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/object-updater/0.log" Oct 04 07:41:17 crc kubenswrapper[4987]: I1004 07:41:17.152805 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/rsync/0.log" Oct 04 07:41:17 crc kubenswrapper[4987]: I1004 07:41:17.210607 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_70e18ecc-8da8-4423-8eff-39f07bf2383f/swift-recon-cron/0.log" Oct 04 07:41:21 crc kubenswrapper[4987]: I1004 07:41:21.034003 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-llv7g"] Oct 04 07:41:21 crc kubenswrapper[4987]: I1004 07:41:21.040249 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-llv7g"] Oct 04 07:41:22 crc kubenswrapper[4987]: I1004 07:41:22.064910 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47de6379-1946-4012-afab-f15535d50880" path="/var/lib/kubelet/pods/47de6379-1946-4012-afab-f15535d50880/volumes" Oct 04 07:41:26 crc kubenswrapper[4987]: I1004 07:41:26.053833 4987 scope.go:117] "RemoveContainer" containerID="c9dda04ae178f74a186ceceaf85fd4f838fcffed0e168fc9f17e8677a8b7f18d" Oct 04 07:41:26 crc kubenswrapper[4987]: I1004 07:41:26.280027 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" event={"ID":"a89a1560-c018-4df7-92aa-795638d35d94","Type":"ContainerStarted","Data":"8e900e0a3690b5fc83f853c136522d2a765aa87a35ad0608cb4411f3837c2d6b"} Oct 04 07:41:28 crc kubenswrapper[4987]: I1004 07:41:28.853746 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/util/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.090577 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/pull/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.098705 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/util/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.111235 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/pull/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.235290 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/util/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.256457 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/pull/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.285001 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2cb2dn_92de2dd9-3957-4f74-a21e-ed5b040b5de4/extract/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.428979 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/extract-utilities/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.566361 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/extract-content/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.566369 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/extract-content/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.579137 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/extract-utilities/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.747770 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/extract-utilities/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.760097 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/extract-content/0.log" Oct 04 07:41:29 crc kubenswrapper[4987]: I1004 07:41:29.979675 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/extract-utilities/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.206501 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/extract-utilities/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.268333 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/extract-content/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.281568 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/extract-content/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.287701 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rmvdf_cb226448-ba0b-4cf6-9000-1f3390408256/registry-server/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.410285 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/extract-utilities/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.530995 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/extract-content/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.637226 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-r46zk_4d6cbfa1-a2bb-4c11-87b2-a5897573ed88/marketplace-operator/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.841130 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/extract-utilities/0.log" Oct 04 07:41:30 crc kubenswrapper[4987]: I1004 07:41:30.950122 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-v9n7c_281f644d-ab95-4743-918b-660dd2644aa6/registry-server/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.028259 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-9e77-account-create-656md"] Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.028367 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/extract-utilities/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.041575 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-9e77-account-create-656md"] Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.065664 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/extract-content/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.113049 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/extract-content/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.284039 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/extract-utilities/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.287590 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/extract-content/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.380152 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jq5sw_c5c87be9-03bb-486d-8024-8f88ee8f8384/registry-server/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.484972 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/extract-utilities/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.622329 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/extract-utilities/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.629169 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/extract-content/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.631795 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/extract-content/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.788008 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/extract-utilities/0.log" Oct 04 07:41:31 crc kubenswrapper[4987]: I1004 07:41:31.839583 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/extract-content/0.log" Oct 04 07:41:32 crc kubenswrapper[4987]: I1004 07:41:32.061783 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2559bd-1713-4220-b9d2-41259f9b5d54" path="/var/lib/kubelet/pods/cc2559bd-1713-4220-b9d2-41259f9b5d54/volumes" Oct 04 07:41:32 crc kubenswrapper[4987]: I1004 07:41:32.332368 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rj2zj_1deff79f-24f7-42c9-9485-b846b87ada21/registry-server/0.log" Oct 04 07:41:38 crc kubenswrapper[4987]: I1004 07:41:38.033022 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-q4svf"] Oct 04 07:41:38 crc kubenswrapper[4987]: I1004 07:41:38.043773 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-q4svf"] Oct 04 07:41:38 crc kubenswrapper[4987]: I1004 07:41:38.062727 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d30bee49-7223-4652-9615-7155381ad95b" path="/var/lib/kubelet/pods/d30bee49-7223-4652-9615-7155381ad95b/volumes" Oct 04 07:41:48 crc kubenswrapper[4987]: I1004 07:41:48.330796 4987 scope.go:117] "RemoveContainer" containerID="fc568d1505d3d44139b68fc46c9c52ef7bfbfb1f9cbe0fdd2bdc2915c1e8d852" Oct 04 07:41:48 crc kubenswrapper[4987]: I1004 07:41:48.354361 4987 scope.go:117] "RemoveContainer" containerID="d9c1f7f2a5a8d9b56507bbba6bce078c17c6933e8777938d7deca9210e408f75" Oct 04 07:41:48 crc kubenswrapper[4987]: I1004 07:41:48.428346 4987 scope.go:117] "RemoveContainer" containerID="24a15e9b45842d16127933029f731dda4bbf688478307e124e8df77cc136fdb3" Oct 04 07:42:47 crc kubenswrapper[4987]: I1004 07:42:47.004585 4987 generic.go:334] "Generic (PLEG): container finished" podID="b67cc4cf-a982-4378-81cf-e179d0149f84" containerID="0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3" exitCode=0 Oct 04 07:42:47 crc kubenswrapper[4987]: I1004 07:42:47.004760 4987 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" event={"ID":"b67cc4cf-a982-4378-81cf-e179d0149f84","Type":"ContainerDied","Data":"0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3"} Oct 04 07:42:47 crc kubenswrapper[4987]: I1004 07:42:47.005728 4987 scope.go:117] "RemoveContainer" containerID="0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3" Oct 04 07:42:47 crc kubenswrapper[4987]: I1004 07:42:47.733342 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gvjbx_must-gather-nc5f2_b67cc4cf-a982-4378-81cf-e179d0149f84/gather/0.log" Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.325166 4987 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gvjbx/must-gather-nc5f2"] Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.327918 4987 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" podUID="b67cc4cf-a982-4378-81cf-e179d0149f84" containerName="copy" containerID="cri-o://6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4" gracePeriod=2 Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.331204 4987 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gvjbx/must-gather-nc5f2"] Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.738799 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gvjbx_must-gather-nc5f2_b67cc4cf-a982-4378-81cf-e179d0149f84/copy/0.log" Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.739371 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.857904 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b67cc4cf-a982-4378-81cf-e179d0149f84-must-gather-output\") pod \"b67cc4cf-a982-4378-81cf-e179d0149f84\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.858106 4987 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9hw2\" (UniqueName: \"kubernetes.io/projected/b67cc4cf-a982-4378-81cf-e179d0149f84-kube-api-access-r9hw2\") pod \"b67cc4cf-a982-4378-81cf-e179d0149f84\" (UID: \"b67cc4cf-a982-4378-81cf-e179d0149f84\") " Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.882853 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67cc4cf-a982-4378-81cf-e179d0149f84-kube-api-access-r9hw2" (OuterVolumeSpecName: "kube-api-access-r9hw2") pod "b67cc4cf-a982-4378-81cf-e179d0149f84" (UID: "b67cc4cf-a982-4378-81cf-e179d0149f84"). InnerVolumeSpecName "kube-api-access-r9hw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.940764 4987 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b67cc4cf-a982-4378-81cf-e179d0149f84-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b67cc4cf-a982-4378-81cf-e179d0149f84" (UID: "b67cc4cf-a982-4378-81cf-e179d0149f84"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.960472 4987 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b67cc4cf-a982-4378-81cf-e179d0149f84-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 07:42:54 crc kubenswrapper[4987]: I1004 07:42:54.960515 4987 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9hw2\" (UniqueName: \"kubernetes.io/projected/b67cc4cf-a982-4378-81cf-e179d0149f84-kube-api-access-r9hw2\") on node \"crc\" DevicePath \"\"" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.066362 4987 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gvjbx_must-gather-nc5f2_b67cc4cf-a982-4378-81cf-e179d0149f84/copy/0.log" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.066902 4987 generic.go:334] "Generic (PLEG): container finished" podID="b67cc4cf-a982-4378-81cf-e179d0149f84" containerID="6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4" exitCode=143 Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.066969 4987 scope.go:117] "RemoveContainer" containerID="6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.066967 4987 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gvjbx/must-gather-nc5f2" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.089788 4987 scope.go:117] "RemoveContainer" containerID="0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.136354 4987 scope.go:117] "RemoveContainer" containerID="6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4" Oct 04 07:42:55 crc kubenswrapper[4987]: E1004 07:42:55.136927 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4\": container with ID starting with 6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4 not found: ID does not exist" containerID="6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.136994 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4"} err="failed to get container status \"6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4\": rpc error: code = NotFound desc = could not find container \"6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4\": container with ID starting with 6551620df21a6217c7eea37ba12620fab7a2d7400fe65712f887686ca43cd1f4 not found: ID does not exist" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.137033 4987 scope.go:117] "RemoveContainer" containerID="0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3" Oct 04 07:42:55 crc kubenswrapper[4987]: E1004 07:42:55.137434 4987 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3\": container with ID starting with 0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3 not found: ID does not exist" containerID="0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3" Oct 04 07:42:55 crc kubenswrapper[4987]: I1004 07:42:55.137473 4987 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3"} err="failed to get container status \"0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3\": rpc error: code = NotFound desc = could not find container \"0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3\": container with ID starting with 0fd2eb25da91161a663b71541fdc5f2e78b90dbf84d74ab75b028462cd3883a3 not found: ID does not exist" Oct 04 07:42:56 crc kubenswrapper[4987]: I1004 07:42:56.064292 4987 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67cc4cf-a982-4378-81cf-e179d0149f84" path="/var/lib/kubelet/pods/b67cc4cf-a982-4378-81cf-e179d0149f84/volumes" Oct 04 07:43:42 crc kubenswrapper[4987]: I1004 07:43:42.703434 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:43:42 crc kubenswrapper[4987]: I1004 07:43:42.704190 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 07:44:12 crc kubenswrapper[4987]: I1004 07:44:12.703773 4987 patch_prober.go:28] interesting pod/machine-config-daemon-dvv6j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 07:44:12 crc kubenswrapper[4987]: I1004 07:44:12.704855 4987 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-dvv6j" podUID="a89a1560-c018-4df7-92aa-795638d35d94" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070150123024437 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070150124017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070143615016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070143616015460 5ustar corecore